No project description provided
Project description
Implementation of https://arxiv.org/abs/1904.00962 for large batch, large learning rate training.
The paper doesn't specify clamp values for ϕ, so I use 10.
Bonus: TensorboardX logging (example below).
Try the sample
git clone git@github.com:cybertronai/pytorch-lamb.git
cd pytorch-lamb
pip install -e .
python test_lamb.py
tensorboard --logdir=runs
Sample results
At --lr=.02
, the Adam optimizer is unable to train.
Red: python test_lamb.py --batch-size=512 --lr=.02 --wd=.01 --log-interval=30 --optimizer=adam
Blue: python test_lamb.py --batch-size=512 --lr=.02 --wd=.01 --log-interval=30 --optimizer=lamb
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
pytorch_lamb-1.0.0.tar.gz
(3.1 kB
view hashes)
Built Distribution
Close
Hashes for pytorch_lamb-1.0.0-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 0cdc4dffe0b402ee55f482466ebab5b25332a1036cef595d5c733bac0a37be71 |
|
MD5 | 0a4e996785f37d35ddb651714c98692d |
|
BLAKE2b-256 | 43983bce14a319317a2856db722f2542d329baf42845fa53563d0d749c5a2d40 |