Skip to main content

A framework to define a machine learning pipeline

Project description

mlpipeline

This is a simple framework to organize you machine learning workflow. It automates most of the basic functionalities such as logging, a framework for testing models and gluing together different steps at different stages. This project came about as a result of me abstracting the boilerplate code and automating different parts of the process.

The aim of this simple framework is to consolidate the different sub-problems (such as loading data, model configurations, training process, evaluation process, exporting trained models, etc.) when working/researching with machine learning models. This allows the user to define how the different sub-problems are to be solved using their choice of tools and mlpipeline would handle piecing them together.

Core operations

This framework chains the different operations (sub-problems) depending on the mode it is executed in. mlpipeline currently has 3 modes:

  • TEST mode: When in TEST mode, it doesn't perform any logging or tracking. It creates a temporary empty directory for the experiment to store the artifacts of an experiment in. When developing and testing the different operations, this mode can be used.
  • RUN mode: In this mode, logging and tracking is performed. In addition, for each experiment run (referred to as a experiment version in mlpipeline) a directory is created for artifacts to be stored.
  • EXPORT mode: In this mode, the exporting related operations will be executed instead of the training/evaluation related operations.

In addition to providing different modes, the pipeline also supports logging and recording various details. Currently mlpipeline records all logs, metrics and artifacts using a basic log files as well using mlflow <https://github.com/databricks/mlflow>_.

The following information is recorded:

  • The scripts that were executed/imported in relation to an experiment.
  • The any output results
  • The metrics and parameters

Documentation

The documentation is hosted at ReadTheDocs <https://mlpipeline.readthedocs.io/>_.

Installing

Can be installed directly using the Python Package Index using pip::

pip install mlpipeline

Usage

work in progress

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

mlpipeline-2.0a4.post13.tar.gz (24.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

mlpipeline-2.0a4.post13-py3-none-any.whl (29.7 kB view details)

Uploaded Python 3

File details

Details for the file mlpipeline-2.0a4.post13.tar.gz.

File metadata

  • Download URL: mlpipeline-2.0a4.post13.tar.gz
  • Upload date:
  • Size: 24.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.20.1 setuptools/41.0.1 requests-toolbelt/0.9.1 tqdm/4.30.0 CPython/3.7.3

File hashes

Hashes for mlpipeline-2.0a4.post13.tar.gz
Algorithm Hash digest
SHA256 714e4b80d2bcf722b68ae2e9026615cc8792f50bd7511894fe5b6eae848276d3
MD5 fb203efe1fb93cac90ea6b1e1981f0d0
BLAKE2b-256 413d1c97276ab3681abe1b4383036d144aa05109e38ecd68af2266c7c6de74ac

See more details on using hashes here.

File details

Details for the file mlpipeline-2.0a4.post13-py3-none-any.whl.

File metadata

  • Download URL: mlpipeline-2.0a4.post13-py3-none-any.whl
  • Upload date:
  • Size: 29.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.20.1 setuptools/41.0.1 requests-toolbelt/0.9.1 tqdm/4.30.0 CPython/3.7.3

File hashes

Hashes for mlpipeline-2.0a4.post13-py3-none-any.whl
Algorithm Hash digest
SHA256 62ec90fd12b70df8804a20fa631fdaab8a45c41ef76654ec0b3e694ae9745c33
MD5 30d0827807428d865ec2b5bab7f92465
BLAKE2b-256 8c818e7a9e9314908faea7ee04e74f284cec2dbaa94352dac35e72615489960f

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page