Skip to main content

Tracking and config of machine learning runs

Project description

Build Status License PyPI version fury.io Documentation Code style: black GitHub commit activity GitHub release (latest SemVer) Join MLOps Live

MLRun logo

MLRun - The Open Source MLOps Orchestration Framework

MLRun enables production pipeline design using a modular strategy, where the different parts contribute to a continuous, automated, and far simpler path from research and development to scalable production pipelines, without refactoring code, adding glue logic, or spending significant efforts on data and ML engineering.

MLRun uses Serverless Function technology: write the code once, using your preferred development environment and simple “local” semantics, and then run it as-is on different platforms and at scale. MLRun automates the build process, execution, data movement, scaling, versioning, parameterization, outputs tracking, CI/CD integration, deployment to production, monitoring, and more.

Those easily developed data or ML “functions” can then be published or loaded from a marketplace and used later to form offline or real-time production pipelines with minimal engineering efforts.

mlrun-flow


Data preparation, model development, model and application delivery, and end to end monitoring are tightly connected: they cannot be managed in silos. This is where MLRun MLOps orchestration comes in. ML, data, and DevOps/MLOps teams collaborate using the same set of tools, practices, APIs, metadata, and version control.

MLRun simplifies & accelerates the time to production.

Architecture

pipeline

MLRun is composed of the following layers:

  • Feature Store — collects, prepares, catalogs, and serves data features for development (offline) and real-time (online) usage for real-time and batch data. See also Feature store: data ingestion and Feature store: data retrieval, as well as the Feature Store tutorials.
  • ML CI/CD pipeline — automatically trains, tests, optimizes, and deploys or updates models using a snapshot of the production data (generated by the feature store) and code from the source control (Git).
  • Real-Time Serving Pipeline — Rapid deployment of scalable data and ML pipelines using real-time serverless technology, including the API handling, data preparation/enrichment, model serving, ensembles, driving and measuring actions, etc.
  • Real-Time monitoring and retraining — monitors data, models, and production components and provides a feedback loop for exploring production data, identifying drift, alerting on anomalies or data quality issues, triggering re-training jobs, measuring business impact, etc.

Get started

It's easy to start using MLRun:

  1. Install the MLRun service locally using Docker or over Kubernetes Cluster. Alternatively, you can use Iguazio's managed MLRun service
  2. Set up your client environment to work with the service.
  3. Follow the Quick Start tutorial and Additional Tutorials and Examples to learn how to use MLRun to develop and deploy machine learning applications to production.

For hands-on learning, try the MLRun Katakoda Scenarios. And you can watch the Tutorial on Youtube to see the flow in action.

MLRun documentation

Read more in the MLRun documentation, including:

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

mlrun-1.1.1rc5-py3-none-any.whl (1.0 MB view details)

Uploaded Python 3

File details

Details for the file mlrun-1.1.1rc5-py3-none-any.whl.

File metadata

  • Download URL: mlrun-1.1.1rc5-py3-none-any.whl
  • Upload date:
  • Size: 1.0 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.7.13

File hashes

Hashes for mlrun-1.1.1rc5-py3-none-any.whl
Algorithm Hash digest
SHA256 e9fe860b9fc5448d59200e5f0d176ba2fbeb304902e5f243761c1e688dc49e1f
MD5 d42829ad5c9329ae2cdc04214eadf9b4
BLAKE2b-256 482ba096d8e36f116c53b0ac1d26199f45c94d3381276837ed977bcb55465f08

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page