Skip to main content

MLflow is an open source platform for the complete machine learning lifecycle

Project description

📣 This is the mlflow-skinny package, a lightweight MLflow package without SQL storage, server, UI, or data science dependencies. Additional dependencies can be installed to leverage the full feature set of MLflow. For example:

  • To use the mlflow.sklearn component of MLflow Models, install scikit-learn, numpy and pandas.
  • To use SQL-based metadata storage, install sqlalchemy, alembic, and sqlparse.
  • To use serving-based features, install flask and pandas.

Note: When using mlflow-skinny, set the tracking URI to your remote MLflow server:

export MLFLOW_TRACKING_URI="http://your-mlflow-server:5000"



MLflow logo

The Open Source AI Engineering Platform for Agents, LLMs & Models

MLflow is the largest open source AI engineering platform for agents, LLMs, and ML models. MLflow enables teams of all sizes to debug, evaluate, monitor, and optimize production-quality AI applications while controlling costs and managing access to models and data. With over 60 million monthly downloads, thousands of organizations rely on MLflow each day to ship AI to production with confidence.

MLflow's comprehensive feature set for agents and LLM applications includes production-grade observability, evaluation, prompt management, prompt optimization and an AI Gateway for managing costs and model access. Learn more at MLflow for LLMs and Agents.


Get Started in 3 Simple Steps

From zero to full-stack LLMOps in minutes. No complex setup or major code changes required. Get Started →

1. Start MLflow Server

uvx mlflow server

2. Enable Logging

import mlflow

mlflow.set_tracking_uri("http://localhost:5000")
mlflow.openai.autolog()

3. Run Your Code

from openai import OpenAI

client = OpenAI()
client.responses.create(
    model="gpt-5.4-mini",
    input="Hello!",
)

Explore traces and metrics in the MLflow UI at http://localhost:5000.

LLMs & Agents

MLflow provides everything you need to build, debug, evaluate, and deploy production-quality LLM applications and AI agents. Supports Python, TypeScript/JavaScript, Java and any other programming language. MLflow also natively integrates with OpenTelemetry and MCP.

Observability

Observability

Capture complete traces of your LLM applications and agents for deep behavioral insights. Built on OpenTelemetry, supporting any LLM provider and agent framework. Monitor production quality, costs, and safety.

Getting Started →

Evaluation

Evaluation

Run systematic evaluations, track quality metrics over time, and catch regressions before they reach production. Choose from 50+ built-in metrics and LLM judges, or define your own.

Getting Started →

Prompts & Optimization

Prompts & Optimization

Version, test, and deploy prompts with full lineage tracking. Automatically optimize prompts with state-of-the-art algorithms to improve performance.

Getting Started →

AI Gateway

AI Gateway

Unified API gateway for all LLM providers. Route requests, manage rate limits, handle fallbacks, and control costs through an OpenAI-compatible interface with built-in credential management, guardrails and traffic splitting for A/B testing.

Getting Started →

Model Training

For machine learning and deep learning model development, MLflow provides a full suite of tools to manage the ML lifecycle:

  • Experiment Tracking — Track models, parameters, metrics, and evaluation results across experiments
  • Model Evaluation — Automated evaluation tools integrated with experiment tracking
  • Model Registry — Collaboratively manage the full lifecycle of ML models
  • Deployment — Deploy models to batch and real-time scoring on Docker, Kubernetes, Azure ML, AWS SageMaker, and more

Learn more at MLflow for Model Training.

Integrations

MLflow supports all agent frameworks, LLM providers, tools, and programming languages. We offer one-line automatic tracing for more than 60 frameworks. See the full integrations list.

OpenTelemetry


OpenTelemetry

Agent Frameworks (Python)


LangChain

LangGraph

OpenAI Agent

DSPy

PydanticAI

Google ADK

Microsoft Agent

CrewAI

LlamaIndex

AutoGen

Strands

LiveKit Agents

Agno

Bedrock AgentCore

Smolagents

Semantic Kernel

DeepAgent

AG2

Haystack

Koog

txtai

Pipecat

Watsonx

Agent Frameworks (TypeScript)


LangChain

LangGraph

Vercel AI SDK

Mastra

VoltAgent

Agent Frameworks (Java)


Spring AI

Quarkus LangChain4j

Model Providers


OpenAI

Anthropic

Databricks

Gemini

Amazon Bedrock

LiteLLM

Mistral

xAI / Grok

Ollama

Groq

DeepSeek

Qwen

Moonshot AI

Cohere

BytePlus

Novita AI

FireworksAI

Together AI

Gateways


Databricks

LiteLLM Proxy

Vercel AI Gateway

OpenRouter

Portkey

Helicone

Kong AI Gateway

PydanticAI Gateway

TrueFoundry

Tools & No-Code


Instructor

Claude Code

Opencode

Langfuse

Arize / Phoenix

Goose

Langflow

Hosting MLflow

MLflow can be used in a variety of environments, including your local environment, on-premises clusters, cloud platforms, and managed services. Being an open-source platform, MLflow is vendor-neutral — whether you're building AI agents, LLM applications, or ML models, you have access to MLflow's core capabilities.


Databricks

Amazon SageMaker

Azure ML

Nebius

Self-Hosted

💭 Support

  • For help or questions about MLflow usage (e.g. "how do I do X?") visit the documentation.
  • In the documentation, you can ask the question to our AI-powered chat bot. Click on the "Ask AI" button at the right bottom.
  • Join the virtual events like office hours and meetups.
  • To report a bug, file a documentation issue, or submit a feature request, please open a GitHub issue.
  • For release announcements and other discussions, please subscribe to our mailing list (mlflow-users@googlegroups.com) or join us on Slack.

🤝 Contributing

We happily welcome contributions to MLflow!

Please see our contribution guide to learn more about contributing to MLflow.

⭐️ Star History

Star History Chart

✏️ Citation

If you use MLflow in your research, please cite it using the "Cite this repository" button at the top of the GitHub repository page, which will provide you with citation formats including APA and BibTeX.

👥 Core Members

MLflow is currently maintained by the following core members with significant contributions from hundreds of exceptionally talented community members.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

mlflow_skinny-3.11.0.tar.gz (2.6 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

mlflow_skinny-3.11.0-py3-none-any.whl (3.2 MB view details)

Uploaded Python 3

File details

Details for the file mlflow_skinny-3.11.0.tar.gz.

File metadata

  • Download URL: mlflow_skinny-3.11.0.tar.gz
  • Upload date:
  • Size: 2.6 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for mlflow_skinny-3.11.0.tar.gz
Algorithm Hash digest
SHA256 723c790891db42d4f6e274d92fe207dd11f9712ef8956888a95b40d27f69598a
MD5 43caa0913212aabc5dc690321c22ed05
BLAKE2b-256 f8a6c69b771a20c00d6c5d2539a4d62300fbdb535c8dde194331b4b2b7c31239

See more details on using hashes here.

Provenance

The following attestation bundles were made for mlflow_skinny-3.11.0.tar.gz:

Publisher: python.yml on mlflow/releases

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file mlflow_skinny-3.11.0-py3-none-any.whl.

File metadata

  • Download URL: mlflow_skinny-3.11.0-py3-none-any.whl
  • Upload date:
  • Size: 3.2 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for mlflow_skinny-3.11.0-py3-none-any.whl
Algorithm Hash digest
SHA256 0f9b788cadce5e68c6ac1c3fc368e35149a8ed013970195021e873f6d0a36180
MD5 8bab46a382efc0165fcb4dd3027a337f
BLAKE2b-256 b1dc5d0f9c156ca5d08870cacc3763d2136d2a6e73b3e828a5e2bdba0bf297fe

See more details on using hashes here.

Provenance

The following attestation bundles were made for mlflow_skinny-3.11.0-py3-none-any.whl:

Publisher: python.yml on mlflow/releases

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page