Skip to main content

Complete Machine Learning Framework for Production

Project description

FRAMEWORM

Production-Grade Generative AI Framework

PyPI Python Tests Coverage Documentation License


The Product Family

FRAMEWORM is not a single tool. It is four products that cover the full ML lifecycle — from training to production.

FRAMEWORM          → core ML infrastructure
FRAMEWORM AGENT    → autonomous training monitor
FRAMEWORM SHIFT    → distribution drift detection
FRAMEWORM DEPLOY   → one-command model deployment

FRAMEWORM (Core)

Production-grade ML infrastructure. Dependency graph engine, Bayesian HPO, mixed precision training, TorchScript + ONNX export, and 6 built-in generative architectures.

Built-in architectures: VAE · DCGAN · DDPM · VQ-VAE-2 · ViT-GAN · CFG-DDPM

pip install frameworm
frameworm init my-project
cd my-project
frameworm train --config config.yaml
Feature Details
Dependency graph engine Kahn's topological sort, DFS cycle detection, parallel execution
Experiment tracking Git hash + dataset checksum snapshots, SQLite backend
Hyperparameter search Grid, random, Bayesian — no external server needed
Config system YAML inheritance: base.yaml → model.yaml → experiment.yaml
Plugin system Drop in new architectures without touching core
Export TorchScript + ONNX, optional INT8 quantization

FRAMEWORM AGENT

Autonomous training monitor. Watches loss curves and gradient norms in real time, detects anomalies, and intervenes automatically.

frameworm agent start --experiment my_run
Anomaly Detection Agent Action
Gradient explosion Grad norm threshold Cut LR 95%, reinit if dead
Loss divergence Rolling window comparison Reduce LR 85%
Training plateau Near-zero variance at high loss Full model reinit + healthy LR
Oscillation Coefficient of variation Reduce LR 90%
Loss spike Statistical outlier Watch mode

Benchmark — 10 Experiments

Each run paired with a shadow run (no agent) as counterfactual comparison.

#   Experiment           Condition        Agent Loss   Shadow Loss    D Loss   Resolved
─────────────────────────────────────────────────────────────────────────────────────────
1   Baseline-1           baseline            70.39        70.37        -0.02       -
2   Baseline-2           baseline            72.61        70.35        -2.27       -
3   Baseline-3           baseline            72.48        70.39        -2.09       -
4   HighLR-1             lr=8e-02            75.71      9999.00     +9923.29      YES
5   HighLR-2             lr=5e-02            69.99        70.37        +0.37      YES
6   GradExplosion-1      grad_explosion      75.93        70.35        -5.58      YES
7   GradExplosion-2      grad_explosion      76.07        70.35        -5.72      YES
8   LowLR-Plateau-1      lr=1e-06            94.65       283.47      +188.82      NO*
9   LowLR-Plateau-2      lr=5e-07            95.31       285.57      +190.26      NO*
10  AgentIntervenes      lr=1e-01            75.84      9999.00     +9923.16      YES
─────────────────────────────────────────────────────────────────────────────────────────
Baseline avg loss:              71.83
Runs with anomalies detected:   7 / 10
Anomalies resolved by agent:    5 / 7
Resolution rate:                71.4%
False positives on clean runs:  0 / 3

Key results:

  • Experiments 4 and 10: agent loss ~75 vs shadow loss 9999 — runs that would have fully diverged were saved
  • Experiments 6 and 7: gradient explosion detected and recovered, final loss within 8% of healthy baseline
  • 0 false positives on clean baseline runs

*Plateau experiments (8, 9): agent detected and intervened 16 times each, reducing loss from ~285 to ~95 (+190 improvement). Full convergence within epoch budget is a known limitation and direction for future work.


FRAMEWORM SHIFT

Drop-in distribution drift detection. Two lines to attach to any model in production.

from frameworm.shift import ShiftMonitor

monitor = ShiftMonitor.from_reference("experiments/my_model.shift")
monitor.check(incoming_batch)   # fires alert automatically if drift detected

Three integration surfaces:

# SDK
monitor = ShiftMonitor("my_model")
monitor.check(batch)

# FastAPI middleware — zero changes to your endpoint
app.add_middleware(ShiftMiddleware, reference="my_model")

# CLI
frameworm shift check --reference train.csv --current live.csv
frameworm shift report --reference train.csv --current live.csv --output report.html
Feature Type Test Output
Numerical Kolmogorov-Smirnov p-value + severity
High-dimensional Maximum Mean Discrepancy drift score
Categorical Chi-squared p-value + severity

Severity levels: NONE · LOW · MEDIUM · HIGH Alerts via: Slack · webhook · log file · stdout


FRAMEWORM DEPLOY

One command from trained model to production API.

frameworm deploy start --model experiments/checkpoints/best.pt --name my_model

What that single command does:

  • Exports model to TorchScript + ONNX
  • Generates a model-aware FastAPI server (correct schema per architecture)
  • Builds a multi-stage Docker image with HEALTHCHECK baked in
  • Starts p50/p95/p99 latency tracking on every request
  • Auto-attaches FRAMEWORM SHIFT drift monitoring
  • Starts a background rollback controller

Auto-rollback: p95 latency spike for 3 consecutive checks → stops bad container, starts previous known-good version, fires Slack alert. No human needed.

Every deployed model exposes:

GET /predict   → inference
GET /health    → liveness
GET /ready     → readiness
GET /metrics   → p50/p95/p99 + error rate
frameworm deploy stop     --name my_model
frameworm deploy status   --name my_model
frameworm deploy rollback --name my_model
frameworm deploy promote  --name my_model --version v2.0 --stage production

Model-aware server generation. FRAMEWORM DEPLOY knows all 6 built-in architectures and generates architecture-specific inference code. A DDPM server runs a 1000-step denoising loop. A VAE server unpacks three return values. A CFG-DDPM server accepts guidance scale. Generic tools can't do this.


Why FRAMEWORM?

FRAMEWORM PyTorch Lightning BentoML MLflow
Training infrastructure ⚠️
Autonomous training monitor ✅ AGENT
Drift detection ✅ SHIFT ⚠️
One-command deployment ✅ DEPLOY ⚠️
Auto-rollback
Model-aware server generation
Experiment lineage ⚠️
External services required ❌ None ⚠️ ⚠️ ⚠️

Quick Start

pip install frameworm

# Train a model
frameworm train --config config.yaml

# Monitor training autonomously
frameworm agent start --experiment my_run

# Check for drift in production
frameworm shift check --reference train.csv --current live.csv

# Deploy to production
frameworm deploy start --model experiments/checkpoints/best.pt --name my_model

Documentation


License

MIT — see LICENSE


Citation

@software{frameworm2026,
  title  = {FRAMEWORM: Production-Grade Generative AI Framework},
  author = {Aakash Ali},
  year   = {2026},
  url    = {https://github.com/Aakash0440/frameworm}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

frameworm-1.2.1.tar.gz (248.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

frameworm-1.2.1-py3-none-any.whl (305.6 kB view details)

Uploaded Python 3

File details

Details for the file frameworm-1.2.1.tar.gz.

File metadata

  • Download URL: frameworm-1.2.1.tar.gz
  • Upload date:
  • Size: 248.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.2

File hashes

Hashes for frameworm-1.2.1.tar.gz
Algorithm Hash digest
SHA256 2a9e1d299969918d99102459202c4d21ee445bc68ec8329d0c82bfdd185fac20
MD5 b6213dd1f2abb0ad4f60013daa541d2f
BLAKE2b-256 98851268698b1a11c92d433f0b57c2e3d08d8595c160dce7acef23256336e65a

See more details on using hashes here.

File details

Details for the file frameworm-1.2.1-py3-none-any.whl.

File metadata

  • Download URL: frameworm-1.2.1-py3-none-any.whl
  • Upload date:
  • Size: 305.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.2

File hashes

Hashes for frameworm-1.2.1-py3-none-any.whl
Algorithm Hash digest
SHA256 f428d2d2f54c29ee08f05a0fdb74b339fbf5847c5354131c1ac2113d1fa1341b
MD5 31410010032531defa7b7113d1c2b177
BLAKE2b-256 0942fde625182da884bf7b4e9ea8ba4764498cd92c3755f536ac577deb9c3cc8

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page