Skip to main content

MLOps Platform with step-based pipeline execution

Project description

ExpOps

expops is a project-based experiment runner: keep each experiment isolated under a workspace, run pipelines, and save run artifacts (with optional tracking/backends).

Install:

pip install expops

The installed CLI command is expops (alias: expops-platform).

Platform Capabilities

ExpOps provides a comprehensive MLOps platform with the following features:

  • Project-Based Workflow: Each ML project is isolated in its own workspace with independent configurations, dependencies, and artifacts
  • DAG Pipeline Execution: Define complex ML pipelines as directed acyclic graphs (DAGs) using NetworkX, with support for parallel execution, conditional logic, and loops
  • Distributed Computing: Execute pipelines on clusters using Dask (with SLURM support) or run locally with multi-worker parallelism
  • Environment Isolation: Automatic virtual environment management (venv/conda) with separate environments for training and reporting
  • Caching & Reproducibility: Intelligent step-level caching with configurable backends (local filesystem, GCS) and reproducibility guarantees via random seed management
  • Static & Dynamic Reporting: Generate static charts (PNG) and interactive dynamic charts that update in real-time

Quick start (built-in template)

mkdir -p ~/expops-workspace && cd ~/expops-workspace

expops create sklearn-basic --template sklearn-basic
expops run sklearn-basic --local

This creates projects/sklearn-basic/ and runs a minimal scikit-learn example. The config is at projects/sklearn-basic/configs/project_config.yaml.

By default, the template uses a local-first cache backend (however the metrics, charts and the web UI only works with a remote kv backend). To enable cross-process live metrics (web UI) or remote backends, update model.parameters.cache.backend in the project config.

To run on a distributed cluster (e.g. SLURM), a cluster_config.yaml should be added under configs and the --local flag can be removed when running.

Create a project

expops create my-project
expops run my-project

Template projects

Templates are available via expops create --template ...:

  • sklearn-basic: runnable project skeleton (configs/data/models/charts + requirements) that trains a tiny sklearn model and generates basic plots

  • premier-league: more comprehensive ML project predicting results of football matches, contains cluster config and dynamic charts as well

Project Structure

Each project follows a standardized directory structure. Here's what each component does:

Configuration Files

configs/project_config.yaml: Main project configuration file that defines:

  • Metadata: Project name, description, version
  • Environment: Virtual environment settings with separate requirements for training and reporting
  • Reproducibility: Random seed configuration and experiment tracking settings
  • Model Configuration: Framework selection, custom script paths, hyperparameters
  • Pipeline Definition: DAG structure (process_adjlist) and process definitions with dependencies
  • Reporting: Chart definitions (static and dynamic) with probe paths for metrics extraction

configs/cluster_config.yaml: Optional cluster execution configuration:

  • Provider: Cluster provider (e.g., slurm, dask)
  • Workers: Number of worker nodes and resource allocation (cores, memory)
  • Queue Settings: Job queue, walltime, and scheduler configuration

Model Code

models/<model_name>.py: Custom model implementation file containing:

  • Process Definitions: Functions decorated with @process() that define pipeline steps
  • Step Functions: Functions decorated with @step() that perform specific operations (data loading, preprocessing, training, inference)
  • Pipeline Logic: Data transformations, model training, evaluation, and ensemble methods
  • Metrics Logging: Integration with log_metric() for experiment tracking

Chart Generation

charts/plot_metrics.py: Python script for static chart generation:

  • Chart Functions: Functions decorated with @chart() that generate PNG visualizations
  • Metrics Access: Reads metrics from previous pipeline steps via ChartContext
  • Static Output: Produces static image files (e.g., PCA scree plots, metric comparisons, distribution histograms)

charts/plot_metrics.js: JavaScript file for dynamic, interactive charts:

  • Real-time Updates: Charts that update dynamically as metrics are logged during pipeline execution
  • Chart.js Integration: Uses Chart.js library for interactive visualizations
  • Live Metrics: Subscribes to metric streams from multiple pipeline steps (e.g., training loss over epochs)
  • Web UI Integration: Rendered in the web UI for interactive exploration

Dependencies

requirements.txt: Main project dependencies for training and inference:

  • Core ML libraries (scikit-learn, XGBoost, PyTorch, TensorFlow, etc.)
  • Data processing libraries (pandas, numpy)
  • Any other dependencies needed for model execution

charts/requirements.txt: Reporting-specific dependencies:

  • Visualization libraries (matplotlib, seaborn)
  • Minimal dependencies needed only for chart generation
  • Kept separate to reduce overhead in training environments

Data & Artifacts

data/: Directory for input datasets (CSV, Parquet, etc.)

logs/: Execution logs for each pipeline run

keys/: Credentials and authentication files (e.g., firestore.json for GCP integration)

Local Web UI

python -m expops.web.server

Open http://127.0.0.1:8000. Choose a project and Run ID (from the configured KV backend in configs/project_config.yaml). The web UI allows you to:

  • Browse projects and runs
  • View static charts
  • Interact with dynamic charts (real-time metric visualization)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

expops-0.1.7.tar.gz (433.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

expops-0.1.7-py3-none-any.whl (431.9 kB view details)

Uploaded Python 3

File details

Details for the file expops-0.1.7.tar.gz.

File metadata

  • Download URL: expops-0.1.7.tar.gz
  • Upload date:
  • Size: 433.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for expops-0.1.7.tar.gz
Algorithm Hash digest
SHA256 836500fa5f5cca2fd7e5ac105a73c78770cc007928837f7a76583ad482299a2a
MD5 104060b4ffa149bd8b7c6f05ef38db92
BLAKE2b-256 c5c079be4e002878aaabd84916f56ea6b787dd55ba781653d4366980aa7ec585

See more details on using hashes here.

Provenance

The following attestation bundles were made for expops-0.1.7.tar.gz:

Publisher: release.yml on local-minima-lab/mlops-platform

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file expops-0.1.7-py3-none-any.whl.

File metadata

  • Download URL: expops-0.1.7-py3-none-any.whl
  • Upload date:
  • Size: 431.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for expops-0.1.7-py3-none-any.whl
Algorithm Hash digest
SHA256 22bd2764ef214a75c04c070f7c08ebb7d671712ae259a856e0888db153fbadca
MD5 55c4c92d36182ee81316dc751bc75d95
BLAKE2b-256 7f8169598f4a4255a782f3480e08b38bf16f590ea62500b302df780350102c9d

See more details on using hashes here.

Provenance

The following attestation bundles were made for expops-0.1.7-py3-none-any.whl:

Publisher: release.yml on local-minima-lab/mlops-platform

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page