Skip to main content

MLflow integration for Inspect AI

Project description

Inspect MLflow Extension

This repository provides an extension that logs Inspect evaluation runs into MLflow.

The intent is:

  • Inspect handles eval execution, scoring, retries, and task logic.
  • MLflow becomes the post-run system of record for analysis and tracking.
  • You should not need to inspect raw .json / .eval logs for normal workflows.

At runtime, the extension consumes Inspect hook events and writes run data to MLflow.

Install

inspect-mlflow is not currently published on PyPI. Install it from this repository:

uv sync

For local development:

uv sync --extra dev

Quick Start

  1. Set up an MLflow tracking server.

This is standard MLflow setup: you can use either an existing remote server (self-hosted or managed) or run one locally.

MLflow docs: https://mlflow.org/docs/latest/self-hosting/architecture/tracking-server/

For quick local validation of this repo, run this SQLite-based server:

uv run mlflow ui --backend-store-uri sqlite:///mlflow.db --host 127.0.0.1 --port 5000

If your MLflow server requires authentication, configure access in your environment before running Inspect.

  1. Set the tracking URI.

For the local server example above, this will be:

export INSPECT_MLFLOW_TRACKING_URI=http://127.0.0.1:5000

For a remote server, set it to your server URL.

  1. Ensure inspect-mlflow is installed in the same environment where you run Inspect (see Install), then run an eval:
inspect eval path/to/task.py

For local testing / runnable quick-start commands, see examples/README.md. Other environment variables are documented in Configuration.

When inspect-mlflow is installed in the Inspect runtime environment, MLflow logging is enabled by default (INSPECT_MLFLOW_ENABLED=true) and runs are logged automatically. If you need to disable logging, set:

export INSPECT_MLFLOW_ENABLED=false
  1. Open MLflow and review runs.

Note: in MLflow 3.x, to inspect run artifacts, switch experiment type to Machine learning (not GenAI apps & agents) and open a run details page.

Where Data Is Stored

inspect-mlflow sends run data to the MLflow server at INSPECT_MLFLOW_TRACKING_URI. The MLflow server then persists that data using MLflow's standard storage model: a backend store for metadata and an artifact store for files.

  • Run metadata (params, metrics, tags, run state) is stored in the server backend store.
  • Artifacts are stored in the server artifact store.

For the Quick Start local command above (mlflow ui --backend-store-uri sqlite:///mlflow.db ...), this means:

  • Backend store is sqlite:///mlflow.db (saved in mlflow.db).
  • Artifacts are served from the default local artifact location (typically ./mlartifacts).

You can change storage locations by configuring MLflow server options such as --backend-store-uri, --artifacts-destination, and --default-artifact-root. For details, see MLflow docs on Tracking Server architecture, Backend Store, and Artifact Store.

Run Mapping

  • One MLflow run is created per Inspect task execution (eval_id).
  • Multi-task or multi-model runs create multiple MLflow runs.
  • inspect eval-set adds grouping via tag inspect.eval_set_id.

Example eval-set run:

inspect eval-set examples/two_tasks.py \
  --id smoke-eval-set \
  --model openai/gpt-4o-mini \
  --log-dir logs/smoke-eval-set

Configuration

Settings can be provided by:

  • environment variables (INSPECT_MLFLOW_*)
  • per-task metadata overrides (metadata keys prefixed with inspect_mlflow_). Environment variables are the primary configuration path; per-task metadata is for targeted overrides.

The sections below group configuration into:

  • Core settings
  • Logging controls
  • Autolog behavior
  • Task metadata overrides

Core

Variable Default Description
INSPECT_MLFLOW_ENABLED true Enable/disable MLflow logging
INSPECT_MLFLOW_TRACKING_URI - MLflow tracking server URI
INSPECT_MLFLOW_EXPERIMENT auto Experiment name (auto-generated if unset)
INSPECT_MLFLOW_RUN_NAME auto Run name (<task>-<short-model>-<eval_id> unless overridden)
INSPECT_MLFLOW_ACCURACY_SCORER auto Scorer name used for inspect.accuracy/inspect.samples_correct (defaults to first task scorer when available)

Logging

Variable Default Description
INSPECT_MLFLOW_LOG_ARTIFACTS true Log artifacts (tables + Inspect logs)
INSPECT_MLFLOW_LOG_TRACES true Enable sample/event trace logging

Autolog

Variable Default Description
INSPECT_MLFLOW_AUTOLOG_ENABLED true Enable MLflow autolog for LLM libraries
INSPECT_MLFLOW_AUTOLOG_MODELS openai,anthropic,langchain,litellm CSV or JSON array of libraries to autolog

Autolog entries are enabled only when both MLflow flavor support and provider dependencies are available.

Task Metadata Overrides

Use task metadata overrides when one specific task needs behavior different from your global environment-based defaults. Each override key should use the inspect_mlflow_ prefix.

Current support is split by scope:

  • Per-task overrides (applied in on_task_start / task lifecycle): enabled, experiment, run_name, log_artifacts, log_traces, accuracy_scorer
  • Run-level only (read once in on_run_start): tracking_uri, autolog_enabled, autolog_models
@task
def my_task():
    return Task(
        dataset=...,
        solver=...,
        metadata={
            "inspect_mlflow_enabled": False,
            "inspect_mlflow_experiment": "my-custom-experiment",
            "inspect_mlflow_accuracy_scorer": "exact",
        },
    )

If you need different tracking URIs or autolog settings, run those tasks in separate Inspect runs with different environment variables.

Accuracy Semantics

inspect-mlflow computes inspect.accuracy and inspect.samples_correct from one selected scorer per task:

  1. INSPECT_MLFLOW_ACCURACY_SCORER / inspect_mlflow_accuracy_scorer when set.
  2. Otherwise, the first scorer declared on the Inspect task.
  3. Otherwise, if a sample has exactly one score, that score is used.

If no scorer can be selected for a sample, it is excluded from accuracy denominator (inspect.samples_scored). If a task has zero scored samples, inspect.accuracy and inspect.samples_correct are emitted as 0.0, and inspect.samples_scored remains 0.

What Gets Logged

Tags

  • inspect.task
  • inspect.model
  • inspect.status (FINISHED / FAILED)
  • inspect.eval_set_id (when running eval-set)

Parameters

  • eval_id
  • flattened task/eval spec fields (model, solver, dataset, config values)
  • inspect.log_file / inspect.log_file_<N> when log file paths are available

Metrics

  • running: inspect.samples, inspect.samples_scored, inspect.samples_correct, inspect.accuracy
  • final: inspect.samples_total, inspect.samples_scored, inspect.samples_correct, inspect.accuracy
  • sample scores: inspect.sample.<eval_id>.<scorer>
  • task aggregate scores: inspect.task.<task>.<scorer>[.<metric>]
  • token usage: inspect.usage.<model>.<token_field>, inspect.tokens.<token_field>

Artifacts

When INSPECT_MLFLOW_LOG_ARTIFACTS=true:

  • inspect/samples.json
  • inspect/messages.json
  • inspect/sample_scores.json
  • inspect/tasks.json
  • inspect/events.json
  • inspect/model_usage.json
  • inspect/logs/<log-file> (best effort)

Traces

When INSPECT_MLFLOW_LOG_TRACES=true:

  • each sample creates a root trace span with sample input/output
  • each chat message in sample.messages is emitted as a child span (message.<idx>.<role>)
  • model/tool/error events are emitted as child spans
  • large traces are bounded (message/event span caps + text truncation) so MLflow UI stays usable
  • traces are logged to the same experiment as the task run and then linked to that run

Examples

Example workflows are in examples/README.md.

Examples are integration smoke tests for this extension and are based on Inspect examples, not a separate benchmark suite.

Acknowledgement

This project was conducted by the Vector Institute and the National Research Council of Canada, on behalf of the Canadian AI Safety Institute.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

inspect_mlflow_ext-0.1.0.tar.gz (31.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

inspect_mlflow_ext-0.1.0-py3-none-any.whl (31.6 kB view details)

Uploaded Python 3

File details

Details for the file inspect_mlflow_ext-0.1.0.tar.gz.

File metadata

  • Download URL: inspect_mlflow_ext-0.1.0.tar.gz
  • Upload date:
  • Size: 31.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for inspect_mlflow_ext-0.1.0.tar.gz
Algorithm Hash digest
SHA256 3f492fa8ccf7dbec58223ba81f2d23e7435dd8a73a39570dad619ac7bdc0b15d
MD5 b5d35eec9431ed2a344b156e0febf3f8
BLAKE2b-256 485942fd717f4f1dffa1093f5d827a16ca6a93f54835464d1e1e70b2fccd11b4

See more details on using hashes here.

File details

Details for the file inspect_mlflow_ext-0.1.0-py3-none-any.whl.

File metadata

File hashes

Hashes for inspect_mlflow_ext-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 515fb45c7adc16e5dac82ee1b39ad81b91a315dacd2d6f210d3f52df925dd592
MD5 d5734d7ef38773642b2217287a3565ea
BLAKE2b-256 7955d8ff8d786a708f38149381225e0c1b4803a5f1d9ad8760d16f0efd4872f0

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page