Skip to main content

MLOps Platform with step-based pipeline execution

Project description

ExpOps

expops is a project-based experiment runner: keep each experiment isolated under a workspace, run pipelines, and save run artifacts (with optional tracking/backends).

Install:

pip install expops

The installed CLI command is mlops (alias: mlops-platform).

Platform Capabilities

ExpOps provides a comprehensive MLOps platform with the following features:

  • Project-Based Workflow: Each ML project is isolated in its own workspace with independent configurations, dependencies, and artifacts
  • DAG Pipeline Execution: Define complex ML pipelines as directed acyclic graphs (DAGs) using NetworkX, with support for parallel execution, conditional logic, and loops
  • Distributed Computing: Execute pipelines on clusters using Dask (with SLURM support) or run locally with multi-worker parallelism
  • Environment Isolation: Automatic virtual environment management (venv/conda) with separate environments for training and reporting
  • Caching & Reproducibility: Intelligent step-level caching with configurable backends (local filesystem, GCS) and reproducibility guarantees via random seed management
  • Static & Dynamic Reporting: Generate static charts (PNG) from Python scripts and interactive dynamic charts (JavaScript) that update in real-time

Quick start (built-in template)

mkdir -p ~/expops-workspace && cd ~/expops-workspace

mlops create demo --template sklearn-basic
mlops run demo

This creates projects/demo/ and runs a minimal scikit-learn example. The config is at projects/demo/configs/project_config.yaml.

Create a project

mlops create my-project
mlops run my-project

Template projects

Templates are available via mlops create --template ...:

  • sklearn-basic: runnable project skeleton (configs/data/models/charts + requirements) that trains a tiny sklearn model and generates basic plots
    Source: src/mlops/templates/sklearn-basic/

Project Structure

Each project follows a standardized directory structure. Here's what each component does:

Configuration Files

configs/project_config.yaml: Main project configuration file that defines:

  • Metadata: Project name, description, version
  • Environment: Virtual environment settings with separate requirements for training and reporting
  • Reproducibility: Random seed configuration and experiment tracking settings
  • Data Sources: Paths to training and validation datasets
  • Model Configuration: Framework selection, custom script paths, hyperparameters
  • Pipeline Definition: DAG structure (process_adjlist) and process definitions with dependencies
  • Reporting: Chart definitions (static and dynamic) with probe paths for metrics extraction

configs/cluster_config.yaml: Optional cluster execution configuration:

  • Provider: Cluster provider (e.g., slurm, dask)
  • Workers: Number of worker nodes and resource allocation (cores, memory)
  • Queue Settings: Job queue, walltime, and scheduler configuration

Model Code

models/<model_name>.py: Custom model implementation file containing:

  • Process Definitions: Functions decorated with @process() that define pipeline steps
  • Step Functions: Functions decorated with @step() that perform specific operations (data loading, preprocessing, training, inference)
  • Pipeline Logic: Data transformations, model training, evaluation, and ensemble methods
  • Metrics Logging: Integration with log_metric() for experiment tracking

Example: In the premier-league project, models/premier_league_model.py defines processes for feature engineering, preprocessing (linear/NN and XGBoost variants), training multiple model variants (A/B testing), inference, best model selection, and ensemble methods.

Chart Generation

charts/plot_metrics.py: Python script for static chart generation:

  • Chart Functions: Functions decorated with @chart() that generate PNG visualizations
  • Metrics Access: Reads metrics from previous pipeline steps via ChartContext
  • Static Output: Produces static image files (e.g., PCA scree plots, metric comparisons, distribution histograms)
  • Matplotlib Integration: Uses matplotlib for visualization with configurable styling

charts/plot_metrics.js: JavaScript file for dynamic, interactive charts:

  • Real-time Updates: Charts that update dynamically as metrics are logged during pipeline execution
  • Chart.js Integration: Uses Chart.js library for interactive visualizations
  • Live Metrics: Subscribes to metric streams from multiple pipeline steps (e.g., training loss over epochs)
  • Web UI Integration: Rendered in the web UI for interactive exploration

Dependencies

requirements.txt: Main project dependencies for training and inference:

  • Core ML libraries (scikit-learn, XGBoost, PyTorch, TensorFlow, etc.)
  • Data processing libraries (pandas, numpy)
  • Any other dependencies needed for model execution

charts/requirements.txt: Reporting-specific dependencies:

  • Visualization libraries (matplotlib, seaborn)
  • Minimal dependencies needed only for chart generation
  • Kept separate to reduce overhead in training environments

Data & Artifacts

data/: Directory for input datasets (CSV, Parquet, etc.)

logs/: Execution logs for each pipeline run

keys/: Credentials and authentication files (e.g., firestore.json for GCP integration)

Local Web UI

python -m mlops.web.server

Open http://127.0.0.1:8000. Choose a project and Run ID (derived from projects/<id>/artifacts/charts/<run-id>). The web UI allows you to:

  • Browse projects and runs
  • View static charts
  • Interact with dynamic charts (real-time metric visualization)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

expops-0.1.1.tar.gz (1.9 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

expops-0.1.1-py3-none-any.whl (202.6 kB view details)

Uploaded Python 3

File details

Details for the file expops-0.1.1.tar.gz.

File metadata

  • Download URL: expops-0.1.1.tar.gz
  • Upload date:
  • Size: 1.9 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for expops-0.1.1.tar.gz
Algorithm Hash digest
SHA256 1aeff19f877f789f2666490e4829ebd43a40d3aec7bc0bb2b5548007315af0ec
MD5 8e5ad927ea2bf57f67a5c9f262fe7cd0
BLAKE2b-256 e2e0d1c7a35be43cda7821d3075219b4f8ef5680ff18c6b346837a0d4233e91e

See more details on using hashes here.

Provenance

The following attestation bundles were made for expops-0.1.1.tar.gz:

Publisher: release.yml on local-minima-lab/mlops-platform

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file expops-0.1.1-py3-none-any.whl.

File metadata

  • Download URL: expops-0.1.1-py3-none-any.whl
  • Upload date:
  • Size: 202.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for expops-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 53eb30ca89be5699f3008effc40f94b7a28f0080595b01d7f61495db5217aa8c
MD5 abb3abbb10f25db675e8aaa21bbad071
BLAKE2b-256 4dc6600076032e37080b065149610a728c51d1d18baa8fb0e6066799a69e8057

See more details on using hashes here.

Provenance

The following attestation bundles were made for expops-0.1.1-py3-none-any.whl:

Publisher: release.yml on local-minima-lab/mlops-platform

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page