Skip to main content

A centralized pattern for creating Machine Learning pipelines

Project description

GitHub Pages
PyPI

ml_pipeline

A centralized pattern for creating Machine Learning pipelines


Installation

pip install kabbes_ml_pipeline


Description

Contains a host of tools for standardizing the format of machine learning pipelines. Provides methods for querying databases, cleaning data, preprocessing data, model operations, exporting results, and more. Each machine learning project is a "child" of this template, with the ability to overwrite any of the default class attributes/methods.

Code Overview

Read this high-level overview is necessary to understand how the package operates.

Code Hierarchy
Models → Model → Input_Files → Input_File → Features → Feature

Each point in the hierarchy has certain methods and attributes associated with it. These methods give you the functionality for operating the pipeline.

Usage

For more in-depth documentation, read the information provided on the Pages. Or better yet, read the source code.

Initializing a Repo for ml_pipeline

  1. Navigate to a directory in command prompt
cd C:/Path/to/Repo
  1. Call the package's main script
python -m ml_pipeline

Navigating the Menu

  1. Run python main_XYZ.py This opens the Models options screen, along with the options for the "Models" class instance.
  2. To navigate one level down to a Model, select the option "Open Model"
  3. Select a Model from the list.
  4. Now in the Model options screen, press enter to navigate back up to Models.
  5. You can navigate from Models->Model->Input_Files->Input_File->Features->Feature and all the way back up.
  6. Option 1 shows "Open XXXXX" to navigate to the next level down in the tree.
  7. Press enter to exit back up to the previous level.

Query new Raw Data

  1. python main_XYZ.py
  2. At the Models options screen, select "Open Model".
  3. Select any Model from the list (this selection does not matter)
  4. In the Model options screen, select "Open Input Files".
  5. Select the first option.
  6. In the Input Files options screen, select "Open Input File"
  7. Select the Input File for which you would like to query new data.
  8. In the Input File options screen, select "Query from Source Database"

Move Query Staged data to Raw Data

  1. Follow steps (1-7) from "Query new Raw Data" to get to the Input File of interest.
  2. In the Input File options screen, select "Move from Query Staging"

Clean a Raw Dataset

  1. Follow steps (1-7) from "Query new Raw Data" to get to the Input File of interest.
  2. In the Input File options screen, select "Clean Raw Dataset"

Running one Model

  1. python main_XYZ.py
  2. At the Models options screen, select the option for "Open Model"
  3. Select the Model you would like to run
  4. In the Model options screen, select the option for "Run Pipeline"

Running all Models

  1. python main_XYZ.py
  2. At the Models options screen, select the option for "Run Models"

Author(s)

James Kabbes

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

kabbes_ml_pipeline-0.6.0.tar.gz (30.4 kB view hashes)

Uploaded Source

Built Distribution

kabbes_ml_pipeline-0.6.0-py3-none-any.whl (39.1 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page