Mlrun
MLRun is an open-source AI orchestration platform designed to build and manage continuous AI applications throughout their lifecycle. It supports data preparation, model training, deployment, and monitoring, integrating with development and CI/CD environments to automate production data workflows and machine learning pipelines. MLRun enables batch and real-time data processing, tracks data lineage, experiments, and metadata, and supports scalable resource management including containers and GPUs. It deploys real-time serving graphs using Nuclio for scalable inference and handles generative AI tasks such as retrieval-augmented generation (RAG), large language model evaluation, and fine-tuning. The platform provides a Function Hub with pre-built functions for ETL, auto-training with frameworks like Scikit-Learn and XGBoost, batch inference, and Azure AutoML. MLRun supports multi-cloud, on-premises, and hybrid environments, allowing collaboration across data, ML, software, and DevOps teams. It requires deploying several services including the MLRun API, UI, database, and Nuclio for full functionality, with Kubernetes preferred for backend setup.
MLRun is an open-source platform that automates end-to-end machine learning pipelines and real-time model serving with integrated lifecycle management.
Continuous AI Application Development
Building and managing machine learning applications that require automated data preparation, model training, deployment, and monitoring.
Real-Time Model Inference
Deploying scalable real-time inference services using serverless functions integrated with Kubernetes or Docker.
Generative AI Tasks
Handling retrieval-augmented generation, large language model evaluation, and fine-tuning within AI workflows.