Skip to main content

Efficient Optimizers

Project description

heavyball

PyPI version License

High-performance, extensible, chainable optimizers for PyTorch.

Why heavyball

  • Lightning-Fast Training: Batched foreach operations deliver significant speedups on large models.
  • Adaptive & Extensible: Built-in AdamW, RMSprop, Schedule-Free algorithms, and PaLM-inspired schedules.
  • Plug-and-Play: Drop-in replacements for torch.optim with seamless integration.
  • Customizable: Chainable API lets you compose optimizers and transforms (MARS correction, cautious updates, orthogonal updates).
  • Battle-Tested: Extensive benchmarks and real-world examples included.

Key Features

  • Foreach-based optimizers: ForeachAdamW, ForeachRMSprop, ForeachSFAdamW, Muon, ADOPT, MSAM (Momentum SAM), …
  • Schedule-Free optimizers with dynamic learning rate adaptation.
  • Advanced update rules: MARS correction, cautious updates, PaLM beta2 scheduling.
  • Chainable transforms for custom optimization recipes.
  • Comprehensive benchmark suite packaged separately as LightBench (../LightBench).
  • Detailed documentation and example-driven tutorials.

Quickstart

Install:

pip install heavyball

Basic usage:

import torch
from torch import nn
from heavyball import ForeachAdamW

model = nn.Sequential(
    nn.Linear(128, 64), nn.ReLU(), nn.Linear(64, 10)
)
optimizer = ForeachAdamW(model.parameters(), lr=1e-3)

for data, target in dataloader:
    optimizer.zero_grad()
    output = model(data)
    loss = torch.nn.functional.cross_entropy(output, target)
    loss.backward()
    optimizer.step()

Benchmarks

Reproduce benchmarks with LightBench (install it via pip install -e ../LightBench from the repo root):

python3 -m lightbench.run_all_benchmarks --opt ForeachSOAP --opt LaProp --opt AdamW --opt Muon --opt ForeachCachedNewtonPSGD  --opt RMSprop --opt OrthoLaProp --opt ForeachSFAdamW --opt ForeachADOPT --opt LaPropOrtho --opt CachedPSGDKron --opt SignLaProp --opt ForeachSOLP --opt PSGDLRA --opt NewtonPSGDLRA --opt NewtonHybrid2PSGDKron --opt NewtonHybrid2PSGDLRA --opt mars-NewtonHybrid2PSGDLRA --opt MSAMLaProp --opt mars-adaptive-NewtonHybrid2PSGDKron  --opt mars-ortho-NewtonHybrid2PSGDKron --opt MuonLaProp --opt mars-unscaled-NewtonHybrid2PSGDKron --opt mars-NewtonHybrid2PSGDKron --opt cautious-AdamW --opt unscaled_cautious-AdamW --opt mars-AdamW  --dtype float32 --steps 1000000 --trials 1000 --parallelism 256 --seeds 1 --difficulties trivial --difficulties easy --difficulties medium --difficulties hard --difficulties extreme --difficulties nightmare --timeout 2880

Migrating from HeavyBall 1.x

  • Read the detailed 2.0.0 migration notes for an end-to-end checklist, including guidance for restoring legacy behaviour when needed.
  • Use scripts/migrate_optimizer_state.py to rewrite pre-2.0 optimizer checkpoints:
    python scripts/migrate_optimizer_state.py path/to/checkpoint.pt heavyball.ForeachAdamW --state-key optimizer
    
    The utility renames legacy state entries, fans them out per parameter view, and injects the HeavyBall metadata block expected by 2.0.0.

Contributing

We welcome contributions! Please check the issue tracker and follow these steps:

  1. Fork the repo and create a feature branch.
  2. Install dev dependencies: pip install -e .[dev].
  3. Run tests: pytest.
  4. Submit a pull request.

License

BSD 3-Clause — see the LICENSE file.


Made by the HeavyBall team.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

heavyball-2.2.1.tar.gz (77.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

heavyball-2.2.1-py3-none-any.whl (53.9 kB view details)

Uploaded Python 3

File details

Details for the file heavyball-2.2.1.tar.gz.

File metadata

  • Download URL: heavyball-2.2.1.tar.gz
  • Upload date:
  • Size: 77.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for heavyball-2.2.1.tar.gz
Algorithm Hash digest
SHA256 ecf50867fa5db12d915c691b9906750cc13a37b8381fb13dd26229768d5929c3
MD5 122feb73d4aecb478df46f11efc3deab
BLAKE2b-256 fb399546f3169e7f8c2f923034247b581b59e1fe0562f80f0d40d929bd0ff9bd

See more details on using hashes here.

Provenance

The following attestation bundles were made for heavyball-2.2.1.tar.gz:

Publisher: release.yaml on HomebrewML/HeavyBall

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file heavyball-2.2.1-py3-none-any.whl.

File metadata

  • Download URL: heavyball-2.2.1-py3-none-any.whl
  • Upload date:
  • Size: 53.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for heavyball-2.2.1-py3-none-any.whl
Algorithm Hash digest
SHA256 bff39760a76660cd2494f0b93f7c12b7f31bd3dca6f0b71f79ca7abde201b944
MD5 3cd16cb6ff55a22f049ea16ac482166d
BLAKE2b-256 6023f1c973d4959f186dc5acc5f7e530fd12d25bfcd985f6e3020ffb0f43a765

See more details on using hashes here.

Provenance

The following attestation bundles were made for heavyball-2.2.1-py3-none-any.whl:

Publisher: release.yaml on HomebrewML/HeavyBall

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page