News

Announced at GTC Paris - #MLRun now integrates with NVIDIA#NeMo microservices

MLRun v1.8 Now Available: Smarter Model Monitoring, Alerts and Tracking

Gilad Shaham | June 5, 2025

We’re proud to announce that the next version of MLRun has been released to community users. On the heels of MLRun v1.7’s focus on monitoring, MLRun v1.8 adds features to make LLM and ML evaluation and monitoring more accessible, practical and resource-efficient.

New Highlights:

  • Smarter LLM model monitoring with in-platform alerts
  • Improved experiment tracking for document-based models
  • Model evaluation before deployment
  • UI improvements with pagination

MLRun is an open-source AI orchestration tool that provides AI practitioners with capabilities to accelerate and streamline the development, deployment and management of gen AI and ML applications. These capabilities include LLM monitoring, fine-tuning, data management, built-in guardrails and more. Ready-made scenarios are available for teams to use instantly, in any environment (both cloud and on-premises).

Now, after months of hard work and collaboration between Iguazio’s engineering team, MLRun users and the broader open-source community, MLRun v1.8 further enhances LLM and ML model monitoring capabilities across the lifecycle.

What is MLRun?

For those who are new to the tool - MLRun is an open-source AI orchestration framework designed to streamline the lifecycle management of ML and generative AI applications, accelerating their path to production. MLRun automates key processes such as data preparation, model tuning, customization, validation and optimization for ML models, LLMs and live AI applications across scalable, elastic infrastructure. With MLRun, organizations can rapidly deploy real-time serving and application pipelines at scale while benefiting from built-in observability and flexible deployment options that support multi-cloud, hybrid and on-premises environments.

Why This Release Focuses on Making Monitoring Easy

AI models and applications are becoming more technologically complex. At the same time, enterprises are realizing their business value and have more aspirational business demands from them than before. However, managing the AI lifecycle, from development to deployment and monitoring, is challenging. Organizations need to overcome issues like taxing resource requirements, technological complexity, rapidly evolving capabilities and practices, the ability to work with and incorporate new models and technologies while keeping up with an ecosystem that is changing fast, filtering out inappropriate content, meeting compliance regulations, and many more. This requires increasingly sophisticated and myriad tools. 

MLRun’s latest updates introduce key enhancements aimed at improving model observability, experiment tracking and evaluation efficiency. These improvements address some of the most common challenges that data scientists and ML engineers face throughout the AI pipeline, such as difficulty detecting toxicity in responses, monitoring resource use, inefficient document-based experiment tracking and compute-heavy model evaluation processes.

V1.8’s enhancements solve real-world challenges in AI and are intended to help you move faster and more effectively while scaling up and reducing costs.

What’s New in MLRun v1.8?

Smarter Model Monitoring with In-Platform Alerts

MLRun now includes monitoring alerts in the UI. These alerts automatically trigger based on predefined severity levels, notifying users of performance degradation, resource spikes, or other anomalies.

From the alerts view, users can dive directly into flagged issues for deeper investigation, instead of relying solely on external systems.

Continuing our emphasis on LLM monitoring from MLRun v1.7, this version provides a seamless monitoring experience inside the platform, reducing the risk of undetected model failures, ensuring more reliable and efficient AI operations.

Improved Experiment Tracking for Documents

MLRun now supports experiment tracking for document-based models, using the LangChain API to integrate directly with vector databases. Users can track documents as artifacts, complete with metadata such as loader type, producer information and collection details.

This makes it easier to version-control and analyze document-related experiments, providing better insights into how different text sources affect model performance and improving model training.

Model Evaluation Before Deployment

MLRun now introduces the ability to monitor and evaluate models before deploying them. Instead of requiring cumbersome and expensive infrastructure setup by the user, MLRun runs the model, returning performance results without consuming unnecessary compute resources.

This enhancement significantly accelerates AI workflows, allowing teams to assess different models efficiently, debug LLMs, enhance fine-tuning, get quick feedback cycles and deploy only the best-performing versions, without waiting for full deployment cycles.

Enhanced UI Experience with Pagination

MLRun now uses pagination in its UI, improving tool responsiveness and reducing page loading times. Users can quickly find relevant experiments, models and monitoring results without excessive scrolling or performance bottlenecks.

With this update, managing large-scale AI projects becomes smoother, allowing teams to focus more on innovation rather than UI frustrations.

Looking Ahead: What’s Next for MLRun

MLRun v1.8 was a continuation of the last version’s focus, expanding LLM monitoring capabilities and updating gen AI application model management and deployment.

We’re looking forward to hearing your feedback about MLRun and your future needs for the upcoming version. So we encourage you to share your insights and requirements in the community.

For a detailed look at the full list of updates, improvements, and to access the new tutorials, visit the MLRun Changelog.