Skip to main content

A base library for ML training (supervised) with environment setup and logging.

Project description

Machine Learning Training Base (ml-training-base)

ml-training-base is a Python package providing base classes and utilities for machine learning projects. Currently, the package only supports supervised learning.

It includes:

  • A configurable logging setup for both console and file outputs.
  • Base classes for data loaders (BaseSupervisedDataLoader).
  • An environment setup class for deterministic training (TrainingEnvironment), ensuring reproducible runs.
  • A base trainer class (BaseSupervisedTrainer) that outlines a typical training workflow in supervised learning.

By using these abstractions, you can quickly spin up a new ML pipeline with consistent structure and easily extend or override specific components to suit your needs.

Table of Contents

  1. Features
  2. Installation
  3. Quick Start
  4. Package Structure
  5. Configuration File
  6. License

Features

  • Reusable Base Classes: Standard building blocks for data loading, training, callbacks, and environment management.
  • Logging Utilities: Automatically configure logging to both console and file, with customizable logging paths.
  • Deterministic Environment Setup: Control Python, NumPy, and TensorFlow seeds for reproducible ML experiments.
  • Clear Project Structure: Easily extend or override abstract methods in your own data loaders, trainers, or environment logic.

Installation

You can install this package locally via:

pip install ml-training-base

Quick Start

  1. Install the package and its dependencies.
  2. Create a YAML configuration file (e.g. config.yaml) with your environment, logging, and data settings.
  3. Import the classes in your script or Jupyter notebook:
import logging
from ml_training_base.data.utils.logging_utils import configure_logger
from ml_training_base.supervised.environments.base_training_environments import KerasTrainingEnvironment
from ml_training_base.supervised.trainers.base_supervised_trainers import BaseSupervisedTrainer
  1. Set up your environment and trainer:
# For example, a custom trainer that inherits from BaseSupervisedTrainer
class MyCustomTrainer(BaseSupervisedTrainer):
    def _setup_model(self):
        # Initialize your model here, e.g., a TensorFlow/Keras or PyTorch model
        pass

    def _build_model(self):
        # Compile or build your model
        pass

    def _setup_callbacks(self):
        # Setup your training callbacks, checkpointing, etc.
        pass

    def _train(self):
        # Implement your training loop or model.fit(...) call
        pass

    def _save_model(self):
        # Save trained model to disk
        pass

    def _evaluate(self):
        # Evaluate your model on the test set
        pass

# Usage:
trainer = MyCustomTrainer(
    config_path="path/to/config.yaml",
    training_env=KerasTrainingEnvironment(logger=logging.getLogger(__name__))
)
trainer.run()

Package Structure

ml-training-base/
├── pyproject.toml
├── src/
│   └── ml_training_base/
│       ├── __init__.py
│       ├── data/
│       │   └── utils/
│       │       ├── __init__.py
│       │       └── logging_utils.py
│       └── supervised/
│           ├── __init__.py
│           ├── environments/
│           │   ├── __init__.py
│           │   └── base_training_environments.py
│           ├── trainers/
│           │   ├── __init__.py
│           │   └── base_supervised_trainers.py
│           └── utils/
│               └── data/
│                   ├── __init__.py
│                   └── base_supervised_data_loader.py
├── tests/
│   ├── __init__.py
│   ├── test_data_loader.py
│   ├── test_environment.py
│   ├── test_logging_utils.py
│   └── test_trainer.py
├── README.md
├── LICENSE
└── pyproject.toml

Key Modules

  • data/utils/logging_utils.py:
    • Contains configure_logger(log_path) utility, which sets up a standardized console and file logger for use throughout the package.
  • supervised/environments/base_training_environments.py:
    • Defines the BaseEnvironment abstract class for handling environment setup.
    • Provides concrete, framework-specific implementations like KerasTrainingEnvironment and PyTorchTrainingEnvironment that manage deterministic setup (setting seeds, configuring hardware options, etc.).
  • supervised/trainers/base_supervised_trainers.py:
    • Contains the core training framework hierarchy.
    • BaseSupervisedTrainer: The framework-agnostic abstract class that defines the training pipeline (run(), _setup_model(), _train(), etc.).
    • BaseKerasSupervisedTrainer & BasePyTorchSupervisedTrainer: Framework-specific abstract classes that implement common boilerplate for Keras (model.fit()) and PyTorch (manual training loop).
  • supervised/utils/data/base_supervised_data_loader.py:
    • Contains the BaseSupervisedDataLoader abstract class. This defines the contract for creating data preparation pipelines (setup_datasets(), get_train_dataset(), etc.) that are used by the trainers.

Configuration File

You can define your runtime settings (e.g., logger paths, environment determinism seeds, model hyperparameters) in a YAML file.

For example:

# Data Configuration and Hyperparameters
data:
  x_data_path: 'data/processed/x_data'
  y_data_path: 'data/processed/y_data'
  logger_path: 'var/log/training.log'
  batch_size: 32
  test_split: 0.1
  validation_split: 0.1

# Model Configuration and Hyperparameters
model:
  attention_dim: 512
  encoder_embedding_dim: 512
  decoder_embedding_dim: 512
  units: 512
  encoder_num_layers: 2
  decoder_num_layers: 4

# Training Configuration and Hyperparameters
training:
  epochs: 100
  early_stop_patience: 5
  weight_decay: null
  dropout_rate: 0.2
  learning_rate: 1e-4

# Environment Configuration
env:
  determinism:
    python_seed: "44478977"
    random_seed: 440651
    numpy_seed: 110789
    tf_seed: 61592

License

This project is licensed under the terms of the MIT License. Feel free to copy, modify, and distribute per its terms.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ml_training_base-0.3.3.tar.gz (17.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

ml_training_base-0.3.3-py3-none-any.whl (18.7 kB view details)

Uploaded Python 3

File details

Details for the file ml_training_base-0.3.3.tar.gz.

File metadata

  • Download URL: ml_training_base-0.3.3.tar.gz
  • Upload date:
  • Size: 17.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.12.8

File hashes

Hashes for ml_training_base-0.3.3.tar.gz
Algorithm Hash digest
SHA256 33a636b2128bdf3b8d9320676f06aaa39aeeed75012e6d1ddd3218d33668ba6c
MD5 807c996390e37ea03686b5ce0296d303
BLAKE2b-256 0bdca38189e9104710a719eac4c13fc0ad451c8482ca09be32b0f2fa72827809

See more details on using hashes here.

Provenance

The following attestation bundles were made for ml_training_base-0.3.3.tar.gz:

Publisher: publish-to-pypi.yml on c-vandenberg/ml-training-base

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file ml_training_base-0.3.3-py3-none-any.whl.

File metadata

File hashes

Hashes for ml_training_base-0.3.3-py3-none-any.whl
Algorithm Hash digest
SHA256 c5292eb7d5de656981628325dcecbadf4718915c588c7de3c4666ae8687444f3
MD5 e90788c8de1d08f9e30f7b99d4dee655
BLAKE2b-256 08e00e1e9a4718fa81f09f80ae67284168fb88aa857b8217bf4ed40445db7b59

See more details on using hashes here.

Provenance

The following attestation bundles were made for ml_training_base-0.3.3-py3-none-any.whl:

Publisher: publish-to-pypi.yml on c-vandenberg/ml-training-base

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page