Skip to main content

Dynamic MLOps Framework with Integrated CLI for Automated ML Project Inception, Kafka-Driven Real-Time Model Monitoring, and Adaptive Canary Deployment Architectures

Project description

MLOPTIFLOW

Dynamic MLOps Framework with Integrated CLI for Automated ML Project Inception, Kafka-Driven Real-Time Model Monitoring, and Adaptive Canary Deployment Architectures

Background

We are introducing a Python framework and library with an integrated CLI, designed to streamline ML lifecycle management by automating project inception, real-time model monitoring, and adaptive canary deployment processes. The library facilitates automated project setup across various configurable ML paradigms (whether it is a Random Forest classification, XGBoost regression, T5-based time-series forecasting, encoder-only / encoder-decoder / decoder-only transformer-based NLP downstream tasks, or practically any other configurable ML paradigm) with optimized directories, subdirectories and configuration files adhering to recommended practices in ML development. The library features a plugin architecture for extensibility, allowing integration with other core components / packages, such as real-time model monitoring with anomaly detection mechanisms, adaptive canary deployment architectures, and integrated UI for monitoring and deployment control. Model monitoring is implemented using high-throughput, low-latency data streaming tool Apache Kafka. Deployed ML models act as Kafka producers, emitting real-time inference data and performance metrics serialized with Apache Avro for schema enforcement and efficiency. Model monitoring is accompanied with anomaly, data, and concept drift detection mechanisms via techniques like PSI, Isolation Forests, or LSTM auto-encoders. Adaptive canary deployment architectures and strategies are implemented specifically for ML models using Kubernetes for container orchestration and Istio as a service mesh for traffic management and routing between baseline and canary versions at granular levels. Integrated UI for monitoring and deployment control is implemented using Vanilla JavaScript and Bootstrap on the client-side, and FastAPI / LitServe on the server-side.

Installation

  1. create a new virtual environment with python ^3.11 and activate it (currently works with virtualenv, venv, and conda)

  2. install poetry:

pip install poetry
  1. install mloptiflow:
pip install mloptiflow
  1. initialize a new project and choose a name and paradigm (currently supported paradigms are: tabular_regression, tabular_classification, demo_tabular_classification)[demo ones are just a minimalistic examples of the paradigm]:
mloptiflow init <your-project-name> --paradigm=<paradigm-name>
  1. cd into your project directory:
cd <your-project-name>
  1. install dependencies:
poetry install

or if using pip:

pip install -r requirements.txt

DEMO Test

  1. create a new virtual environment with python ^3.11 and activate it

  2. install poetry:

pip install poetry
  1. install mloptiflow:
pip install mloptiflow
  1. initialize a new project with the name demo-project and paradigm demo_tabular_classification:
mloptiflow init demo-project --paradigm=demo_tabular_classification
  1. cd into your project directory:
cd demo-project
  1. install dependencies:
poetry install
  1. run the training pipeline:
mloptiflow train start
  1. run and test the inference API:
mloptiflow deploy start --with-api-test

Usage

  • TBA

Support

  • TBA

Roadmap

  • TBA

Contributing

  • do not push directly to the main branch, open MR instead
  • after implementing and before pushing, run:
poetry run python scripts/dev.py lint
poetry run python scripts/dev.py format
poetry run python scripts/dev.py fix
  • after implementing and before pushing, implement corresponding tests and run them:
poetry run pytest
  • if the tests fail, fix them and run tests again until they pass

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

mloptiflow-0.0.72.tar.gz (33.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

mloptiflow-0.0.72-py3-none-any.whl (56.7 kB view details)

Uploaded Python 3

File details

Details for the file mloptiflow-0.0.72.tar.gz.

File metadata

  • Download URL: mloptiflow-0.0.72.tar.gz
  • Upload date:
  • Size: 33.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.7.1 CPython/3.11.4 Linux/5.15.154+

File hashes

Hashes for mloptiflow-0.0.72.tar.gz
Algorithm Hash digest
SHA256 cea0b2ce5f38f802b8261ad30cd8c8dda4ee75db01c54f6be8262e61f60e6be1
MD5 2b6f96c53bc6b30aa4381b62bbb77372
BLAKE2b-256 6a93a5d09de7e801a01241b276b4fd05aecad710e1f6c72092471007919044e9

See more details on using hashes here.

File details

Details for the file mloptiflow-0.0.72-py3-none-any.whl.

File metadata

  • Download URL: mloptiflow-0.0.72-py3-none-any.whl
  • Upload date:
  • Size: 56.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.7.1 CPython/3.11.4 Linux/5.15.154+

File hashes

Hashes for mloptiflow-0.0.72-py3-none-any.whl
Algorithm Hash digest
SHA256 50be339c8bb9b83e675301e027ce2b519fe312cbf2e6275fbdc4a873c9b0d4da
MD5 016c2983499deddf7f06a9954124846f
BLAKE2b-256 e104bb3307d87e7fc10c8cbbc8eb4cfb842fd04e8e4ce9ecfbf22dde697c8f1d

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page