Skip to main content

Reusable deep learning framework core.

Project description

deep-learning-core

Reusable deep learning framework core.

deep-learning-core contains the vendor-neutral training framework that can be reused across many experiment repositories. It is intended to be the public base package, while optional integrations such as Azure are layered on through extras and companion extension packages.

What's New?

  • dl-init is now the primary scaffold command
  • dl-core list makes built-in and local registry discovery easier
  • dl-core add defaults to plain base classes unless --base is given
  • dl-analyze is the primary sweep-analysis CLI
  • dl-analyze now supports explicit ranking metrics and rank methods
  • dl-analyze now persists analysis_cache.json next to sweep_tracking.json
  • dl-analyze now writes versioned reports under analysis/vN.md
  • dl-sync --sweep ... --artifacts now syncs tracked remote artifacts into the local repository when the active backend supports it
  • EMA checkpoints now include a drop-in ema_models_state_dict alongside the normal training weights and EMA resume metadata
  • dl-run --validate-only now performs a real preflight by resolving the configured components without starting training
  • dl-inspect-dataset now summarizes split sizes and one collated batch from the current config
  • dl-smoke now checks one dataset batch and one model forward pass from a config file
  • local artifacts now use:
    • artifacts/runs/<run_name>/...
    • artifacts/sweeps/<sweep_name>/<run_name>/...

Install

Install from PyPI:

pip install deep-learning-core

Install with Azure support:

pip install "deep-learning-core[azure]"

Install with local MLflow support:

pip install "deep-learning-core[mlflow]"

Install with W&B support:

pip install "deep-learning-core[wandb]"

Install with multiple variants:

pip install "deep-learning-core[azure,wandb]"

Install in a uv project:

uv add deep-learning-core

deep-learning-core intentionally ships with the full public runtime dependencies, including torch, torchvision, and opencv-python-headless. The Azure extra pulls in deep-learning-azure, which pins the Azure package versions used by the validated Azure packaging stack. The MLflow extra pulls in deep-learning-mlflow for local MLflow tracking. The W&B extra pulls in deep-learning-wandb and leaves the wandb package itself unpinned.

Package Variants

  • deep-learning-core: local training, local sweeps, local sweep analysis, and the experiment scaffold
  • deep-learning-core[azure]: adds the public dl-azure package for Azure execution and Azure dataset foundations
  • deep-learning-core[mlflow]: adds the public dl-mlflow package for local MLflow integration
  • deep-learning-core[wandb]: adds the public dl-wandb package for Weights & Biases integration

The extension packages stay separate so the base package remains reusable and vendor-neutral.

You can also install the companion packages directly when you want a specific integration without using extras:

pip install deep-learning-azure
pip install deep-learning-mlflow
pip install deep-learning-wandb

Scope

  • Base abstractions and registries
  • Built-in accelerators, callbacks, criterions, metrics, and schedulers
  • The standard trainer and standard dataset flow
  • Built-in augmentations
  • Local execution and sweep orchestration
  • Local sweep analysis from saved artifact summaries
  • Experiment repository scaffolding via dl-init

Out Of Scope

  • Azure ML wiring unless the Azure extra is installed
  • Workspace or datastore conventions
  • Experiment-specific datasets, models, and trainers
  • User-owned configs and private data

Quick Start

uv run dl-core list
uv run dl-init --name my-exp --root-dir .

To initialize the current directory in place, omit --name:

uv run dl-init --root-dir .

The generated experiment repository is the normal consumer entry point. Inside that repository, run uv sync, then run:

uv run dl-run --config configs/base.yaml --validate-only
uv run dl-inspect-dataset --config configs/base.yaml
uv run dl-smoke --config configs/base.yaml
uv run dl-run --config configs/base.yaml
uv run dl-sweep experiments/lr_sweep.yaml --preview
uv run dl-sweep experiments/lr_sweep.yaml --only "*seed_2025*"
uv run dl-sweep experiments/lr_sweep.yaml
uv run dl-analyze --sweep experiments/lr_sweep.yaml
uv run dl-sync --sweep experiments/lr_sweep.yaml --artifacts
uv run dl-analyze --sweep experiments/lr_sweep.yaml --name pareto_eer
uv run dl-analyze --sweep experiments/lr_sweep.yaml --compare latest
uv run dl-analyze --sweep experiments/lr_sweep.yaml --metric test/eer --mode min

New local runs use the flattened artifact layout:

  • artifacts/runs/<run_name>/...
  • artifacts/sweeps/<sweep_name>/<run_name>/...

dl-core does not create a latest symlink for these run directories. Use the concrete run directory names directly.

First Run Workflow

If you are starting from scratch, the minimum path is:

pip install deep-learning-core
uv run dl-init --name my-exp --root-dir .
cd my-exp
uv sync

Then:

  1. open these generated files first:
    • src/datasets/my_exp.py
    • configs/base.yaml
    • scripts/temporary/test_dataset.py
    • scripts/temporary/test_model.py
    • experiments/lr_sweep.yaml
    • AGENTS.md
  2. implement the generated dataset wrapper under src/datasets/my_exp.py
  3. adjust configs/base.yaml so it points at the dataset/model/trainer you want
  4. smoke-check the generated helpers:
uv run python scripts/temporary/test_dataset.py
uv run python scripts/temporary/test_model.py
  1. start with:
uv run dl-run --config configs/base.yaml --validate-only
uv run dl-inspect-dataset --config configs/base.yaml
uv run dl-smoke --config configs/base.yaml
uv run dl-run --config configs/base.yaml

Once that works, move on to:

uv run dl-sweep experiments/lr_sweep.yaml --preview
uv run dl-sweep experiments/lr_sweep.yaml
uv run dl-analyze --sweep experiments/lr_sweep.yaml
uv run dl-analyze --sweep experiments/lr_sweep.yaml \
  --metric test/eer --mode min \
  --metric test/accuracy --mode max \
  --rank-method rank-sum

dl-sweep --preview prints the expanded run matrix without creating configs or starting runs. Use --export sweep_preview.csv or --export sweep_preview.json when you want to save that expansion for review. Use --only and --skip with glob patterns when you want to execute or preview only a subset of generated run names.

dl-inspect-dataset preserves the configured split behavior, but forces single-process loading so you can quickly verify split sizes and inspect one collated batch without starting a trainer.

dl-sync --sweep ... --artifacts syncs tracked run outputs into the local repo. Backends that already write local artifacts simply refresh the tracker paths. Remote-backed integrations can download the run bundle and patch sweep_tracking.json with the resolved local artifact paths.

dl-analyze defaults to ranking by test/accuracy with max. You can make that explicit or override it with one or more --metric / --mode pairs and choose lexicographic, rank-sum, or pareto ranking.

For Azure-backed sweeps, dl-analyze fetches only the requested metric histories instead of downloading every tracked metric history. Those fetched histories are cached in experiments/<sweep_name>/analysis_cache.json. Use --force to ignore and refresh that cache. Reports are written under experiments/<sweep_name>/analysis/ as v1.md, v2.md, and so on unless you pass --name. A matching JSON report is always written next to each Markdown report, and --compare latest or --compare v1 compares the current ranking against a saved report.

EMA Checkpoints

When EMA is enabled with save_in_checkpoint: true, each checkpoint stores:

  • models_state_dict: normal model weights for training resume
  • ema_state_dict: EMA bookkeeping and shadow-parameter state for trainer-side resume
  • ema_models_state_dict: a full drop-in model state dict with EMA parameters and the original model buffers preserved

That means evaluator-side code can load:

  • checkpoint["models_state_dict"]["main"] for normal weights
  • checkpoint["ema_models_state_dict"]["main"] for EMA weights

without needing to reconstruct EMA state manually.

If Azure support is installed, uv run dl-init --with-azure will also scaffold Azure-ready config placeholders and azure-config.json.

If local MLflow support is installed, uv run dl-init --with-mlflow will also scaffold an mlflow callback block and local tracking defaults.

If W&B support is installed, uv run dl-init --with-wandb will also scaffold a wandb callback block, W&B tracking defaults, and .env.example.

Companion Packages

Scaffold Commands

Each dl-core add ... command creates the new module and updates the matching local package __init__.py export list under src/.

uv run dl-core describe ... now also shows a minimal YAML snippet for common config-backed component types such as datasets, models, callbacks, optimizers, and trainers.

Common local component scaffolds:

uv run dl-core add model MyResNet
uv run dl-core add trainer MyTrainer
uv run dl-core add callback MyMetrics
uv run dl-core add metric_manager MyManager
uv run dl-core add sampler MySampler
uv run dl-core add optimizer MyOptimizer
uv run dl-core add scheduler MyScheduler
uv run dl-core add criterion MyLoss
uv run dl-core add augmentation MyAugmentation
uv run dl-core add metric MyMetric
uv run dl-core add executor MyExecutor

Default-base scaffolds for augmentations, metrics, metric managers, criterions, models, and executors now start with ready-to-edit method stubs instead of empty wrapper subclasses.

Sweep scaffolds are supported too:

uv run dl-core add sweep DebugSweep
uv run dl-core add sweep AzureEval --tracking azure_mlflow
uv run dl-core add sweep MlflowBaseline --tracking mlflow
uv run dl-core add sweep WandbAblation --tracking wandb

Generated sweep files:

  • live under experiments/
  • extend ../configs/base_sweep.yaml
  • include runnable defaults in fixed
  • start with grid: {}
  • default the tracker experiment destination to the repository root name unless tracking.experiment_name overrides it
  • let the tracker derive sweep grouping from the filename unless tracking.sweep_name overrides it

Project-specific criterions, optimizers, and schedulers can still be added later with uv run dl-core add ... when they are actually needed.

You can inspect registered components and built-in base classes directly from the CLI:

uv run dl-core list
uv run dl-core list sampler
uv run dl-core list metric_manager --json
uv run dl-core describe dataset my_dataset --root-dir .
uv run dl-core describe model my_resnet --root-dir .
uv run dl-core describe class dl_core.core.FrameWrapper
uv run dl-core describe class dl_azure.datasets.AzureComputeMultiFrameWrapper
uv run dl-core describe dataset my_dataset --root-dir . --json

The built-in sampler list now includes label, which balances samples by any metadata key using either undersample or oversample.

Example sampler config:

dataset:
  sampler:
    label:
      key: attack
      mode: undersample

The describe command shows:

  • resolved class and registered names
  • constructor signature
  • inheritance chain
  • docstring
  • declared properties
  • class-level attributes
  • public methods defined on the class

It does not discover instance attributes created dynamically inside __init__ without constructing the class.

Scaffolds can target a specific base when you need one:

uv run dl-core add dataset MyDataset
uv run dl-core add dataset FrameSet --base frame
uv run dl-core add dataset TextSet --base text_sequence
uv run dl-core add dataset ActSet --base adaptive_computation
uv run dl-core add callback EpochLogger --base metric_logger
uv run dl-core add metric_manager PadMetrics --base standard
uv run dl-core add optimizer AdamwWrapper --base adamw
uv run dl-core add scheduler CosineWrapper --base cosine

Built-in callbacks include dataset_refresh, which rebuilds selected dataset splits at epoch boundaries. Example:

callbacks:
  dataset_refresh:
    refresh_frequency: 1
    splits: [train]

When dl-azure is importable, the dataset scaffold also exposes Azure bases:

uv run dl-core add dataset AzureFrames --base azure_compute_frame
uv run dl-core add dataset AzureSeq --base azure_compute_multiframe
uv run dl-core add dataset AzureStream --base azure_streaming
uv run dl-core add dataset AzureStreamSeq --base azure_streaming_multiframe

Plain deep-learning-core currently exposes dataset bases for:

  • BaseWrapper
  • FrameWrapper
  • TextSequenceWrapper
  • AdaptiveComputationDataset

TextSequenceWrapper adds sequence-aware batch padding for tokenized inputs. AdaptiveComputationDataset adds per-class sample stream helpers for adaptive-time computation trainers. Multiframe dataset bases are still provided through dl-azure.

Releases

  • Publish is the production workflow for PyPI.
  • Trusted publishing is configured through GitHub Actions environments rather than long-lived API tokens.
  • The publish action may upload digital attestations alongside the package. That is expected behavior from pypa/gh-action-pypi-publish.
  • Package metadata keeps runtime dependencies unpinned, so the consuming environment resolves the latest compatible public releases.

Documentation

License

MIT. See LICENSE.

Development Validation

uv run --extra dev pytest
uv run python -m compileall src/dl_core

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

deep_learning_core-0.0.18.tar.gz (468.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

deep_learning_core-0.0.18-py3-none-any.whl (241.5 kB view details)

Uploaded Python 3

File details

Details for the file deep_learning_core-0.0.18.tar.gz.

File metadata

  • Download URL: deep_learning_core-0.0.18.tar.gz
  • Upload date:
  • Size: 468.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for deep_learning_core-0.0.18.tar.gz
Algorithm Hash digest
SHA256 f5c71f9992e85bebbae10746d1522ef6749adca5f466f117ff20952e982c32e7
MD5 5f7f759750a207182e18f18ced495e09
BLAKE2b-256 b517b78b93b746bd85b8b2c671d5559568fab4788aa1dcf91172dc2a15cc4422

See more details on using hashes here.

Provenance

The following attestation bundles were made for deep_learning_core-0.0.18.tar.gz:

Publisher: publish.yml on Blazkowiz47/dl-core

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file deep_learning_core-0.0.18-py3-none-any.whl.

File metadata

File hashes

Hashes for deep_learning_core-0.0.18-py3-none-any.whl
Algorithm Hash digest
SHA256 14c626abfd2529870f54ac9528e90d398118ab2de1d712ba9bc0c4b0832ce5e3
MD5 161973861ecc6b88a595222c7a0bba29
BLAKE2b-256 722c1295f3e4c255f2e8d0d8a08a1cc5bec4cc4aefc3f349b44f950900f7a2bd

See more details on using hashes here.

Provenance

The following attestation bundles were made for deep_learning_core-0.0.18-py3-none-any.whl:

Publisher: publish.yml on Blazkowiz47/dl-core

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page