Skip to main content

MLOps Platform with step-based pipeline execution

Project description

ExpOps

expops is a project-based experiment runner: keep each experiment isolated under a workspace, run pipelines, and save run artifacts (with optional tracking/backends).

Install:

pip install expops

The installed CLI command is expops (alias: expops-platform).

Platform Capabilities

ExpOps provides a comprehensive MLOps platform with the following features:

  • Project-Based Workflow: Each ML project is isolated in its own workspace with independent configurations, dependencies, and artifacts
  • DAG Pipeline Execution: Define complex ML pipelines as directed acyclic graphs (DAGs) using NetworkX, with support for parallel execution, conditional logic, and loops
  • Distributed Computing: Execute pipelines on clusters using Dask (with SLURM support) or run locally with multi-worker parallelism
  • Environment Isolation: Automatic virtual environment management (venv/conda) with separate environments for training and reporting
  • Caching & Reproducibility: Intelligent step-level caching with configurable backends (local filesystem, GCS) and reproducibility guarantees via random seed management
  • Static & Dynamic Reporting: Generate static charts (PNG) and interactive dynamic charts that update in real-time

Quick start (built-in template)

mkdir -p ~/expops-workspace && cd ~/expops-workspace

expops create sklearn-basic --template sklearn-basic
expops run sklearn-basic

This creates projects/sklearn-basic/ and runs a minimal scikit-learn example. The config is at projects/sklearn-basic/configs/project_config.yaml.

By default, the template uses a local-first cache backend (however the metrics, charts and the web UI only works with a remote kv backend). To enable cross-process live metrics (web UI) or remote backends, update model.parameters.cache.backend in the project config.

Create a project

expops create my-project
expops run my-project

Template projects

Templates are available via expops create --template ...:

  • sklearn-basic: runnable project skeleton (configs/data/models/charts + requirements) that trains a tiny sklearn model and generates basic plots

  • premier-league: more comprehensive ML project predicting results of football matches, contains cluster config and dynamic charts as well

Project Structure

Each project follows a standardized directory structure. Here's what each component does:

Configuration Files

configs/project_config.yaml: Main project configuration file that defines:

  • Metadata: Project name, description, version
  • Environment: Virtual environment settings with separate requirements for training and reporting
  • Reproducibility: Random seed configuration and experiment tracking settings
  • Model Configuration: Framework selection, custom script paths, hyperparameters
  • Pipeline Definition: DAG structure (process_adjlist) and process definitions with dependencies
  • Reporting: Chart definitions (static and dynamic) with probe paths for metrics extraction

configs/cluster_config.yaml: Optional cluster execution configuration:

  • Provider: Cluster provider (e.g., slurm, dask)
  • Workers: Number of worker nodes and resource allocation (cores, memory)
  • Queue Settings: Job queue, walltime, and scheduler configuration

Model Code

models/<model_name>.py: Custom model implementation file containing:

  • Process Definitions: Functions decorated with @process() that define pipeline steps
  • Step Functions: Functions decorated with @step() that perform specific operations (data loading, preprocessing, training, inference)
  • Pipeline Logic: Data transformations, model training, evaluation, and ensemble methods
  • Metrics Logging: Integration with log_metric() for experiment tracking

Chart Generation

charts/plot_metrics.py: Python script for static chart generation:

  • Chart Functions: Functions decorated with @chart() that generate PNG visualizations
  • Metrics Access: Reads metrics from previous pipeline steps via ChartContext
  • Static Output: Produces static image files (e.g., PCA scree plots, metric comparisons, distribution histograms)

charts/plot_metrics.js: JavaScript file for dynamic, interactive charts:

  • Real-time Updates: Charts that update dynamically as metrics are logged during pipeline execution
  • Chart.js Integration: Uses Chart.js library for interactive visualizations
  • Live Metrics: Subscribes to metric streams from multiple pipeline steps (e.g., training loss over epochs)
  • Web UI Integration: Rendered in the web UI for interactive exploration

Dependencies

requirements.txt: Main project dependencies for training and inference:

  • Core ML libraries (scikit-learn, XGBoost, PyTorch, TensorFlow, etc.)
  • Data processing libraries (pandas, numpy)
  • Any other dependencies needed for model execution

charts/requirements.txt: Reporting-specific dependencies:

  • Visualization libraries (matplotlib, seaborn)
  • Minimal dependencies needed only for chart generation
  • Kept separate to reduce overhead in training environments

Data & Artifacts

data/: Directory for input datasets (CSV, Parquet, etc.)

logs/: Execution logs for each pipeline run

keys/: Credentials and authentication files (e.g., firestore.json for GCP integration)

Local Web UI

python -m expops.web.server

Open http://127.0.0.1:8000. Choose a project and Run ID (from the configured KV backend in configs/project_config.yaml). The web UI allows you to:

  • Browse projects and runs
  • View static charts
  • Interact with dynamic charts (real-time metric visualization)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

expops-0.1.6.tar.gz (433.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

expops-0.1.6-py3-none-any.whl (431.0 kB view details)

Uploaded Python 3

File details

Details for the file expops-0.1.6.tar.gz.

File metadata

  • Download URL: expops-0.1.6.tar.gz
  • Upload date:
  • Size: 433.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for expops-0.1.6.tar.gz
Algorithm Hash digest
SHA256 74d96d9c29b0fc547dc3a462e668f4986d6f05439f71216e5a598d67c5b31ea2
MD5 5067cb5a219208b55eeed2a91f1d03c5
BLAKE2b-256 7f12b8bda3d5995aa6f3cbe90a3f154067cc438e74f47dd636e764452445f335

See more details on using hashes here.

Provenance

The following attestation bundles were made for expops-0.1.6.tar.gz:

Publisher: release.yml on local-minima-lab/mlops-platform

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file expops-0.1.6-py3-none-any.whl.

File metadata

  • Download URL: expops-0.1.6-py3-none-any.whl
  • Upload date:
  • Size: 431.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for expops-0.1.6-py3-none-any.whl
Algorithm Hash digest
SHA256 5cf1edc98887dd0b01810fe6e89ceaf5b3ad6e26371d00271a3c94b1286921ae
MD5 0ec6ee09c7a2e720c7203a095bc959ef
BLAKE2b-256 0783974ffc689cb9c6f5f43ec58a93a08df686a933d6f319e2a5943191856558

See more details on using hashes here.

Provenance

The following attestation bundles were made for expops-0.1.6-py3-none-any.whl:

Publisher: release.yml on local-minima-lab/mlops-platform

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page