Implementation of the Alberta Plan for AI Research - continual learning with meta-learned step-sizes
Project description
Alberta Framework
A JAX-based research framework implementing components of The Alberta Plan for AI Research in the pursuit of building the foundations of Continual AI.
Warning: This framework is under active research development. The API is unstable and subject to breaking changes between releases. It is not intended for production use.
"The agents are complex only because they interact with a complex world... their initial design is as simple, general, and scalable as possible." — Sutton et al., 2022
Overview
The Alberta Framework provides foundational components for continual reinforcement learning research. Built on JAX for hardware acceleration, the framework emphasizes temporal uniformity every component updates at every time step, with no special training phases or batch processing.
Project Context
This framework is developed as part of my D.Eng. work focusing on the foundations of Continual AI. For more background and context see:
- Research Blog: blog.9600baud.net
- Replicating Sutton '92: The Foundation of Step-size Adaptation
- About the Author: Keith Lawson
Roadmap
Depending on my research trajectory I may or may not implement components required for the plan. The current focus of this framework is the Step 1 Baseline Study, investigating the interaction between adaptive optimizers and online normalization.
| Step | Focus | Status |
|---|---|---|
| 1 | Meta-learned step-sizes (IDBD, Autostep) | Complete |
| 2 | Nonlinear function approximation (MLP, ObGD) | In Progress |
| 3 | GVF predictions, Horde architecture | Planned |
| 4 | Actor-critic with eligibility traces | Planned |
| 5-6 | Off-policy learning, average reward | Planned |
| 7-12 | Hierarchical, multi-agent, world models | Future |
Installation
pip install alberta-framework
# With optional dependencies
pip install alberta-framework[gymnasium] # RL environment support
pip install alberta-framework[dev] # Development (pytest, ruff)
Requirements: Python >= 3.13, JAX >= 0.4, NumPy >= 2.0
Quick Start
import jax.random as jr
from alberta_framework import LinearLearner, IDBD, RandomWalkStream, run_learning_loop
# Non-stationary stream where target weights drift over time
stream = RandomWalkStream(feature_dim=10, drift_rate=0.001)
# Learner with IDBD meta-learned step-sizes
learner = LinearLearner(optimizer=IDBD())
# JIT-compiled training via jax.lax.scan
state, metrics = run_learning_loop(learner, stream, num_steps=10000, key=jr.key(42))
Core Components
Composable Architecture
Learners accept three independent, composable concerns:
- Optimizer — per-weight step-size adaptation (LMS, IDBD, Autostep)
- Bounder — optional global update bounding (ObGDBounding)
- Normalizer — optional online feature normalization (EMANormalizer, WelfordNormalizer)
from alberta_framework import (
LinearLearner, MLPLearner, Autostep, ObGDBounding, EMANormalizer
)
# Linear learner with Autostep + normalization
learner = LinearLearner(
optimizer=Autostep(),
normalizer=EMANormalizer(decay=0.99),
)
# MLP with Autostep + ObGD bounding + normalization
mlp = MLPLearner(
hidden_sizes=(128, 128),
optimizer=Autostep(),
bounder=ObGDBounding(kappa=2.0),
normalizer=EMANormalizer(decay=0.99),
)
Optimizers
Supervised Learning:
- LMS: Fixed step-size baseline
- IDBD: Per-weight adaptive step-sizes via gradient correlation (Sutton, 1992)
- Autostep: Tuning-free adaptation with gradient normalization (Mahmood et al., 2012)
TD Learning:
- TDIDBD: TD learning with per-weight adaptive step-sizes and eligibility traces (Kearney et al., 2019)
- AutoTDIDBD: TD learning with AutoStep-style normalization for improved stability
Bounders
- ObGDBounding: Dynamic update bounding to prevent overshooting (Elsayed et al., 2024). Decoupled from the optimizer so it can be composed with any optimizer.
Normalizers
Online feature normalization for handling varying feature scales:
- EMANormalizer: Exponential moving average — suitable for non-stationary distributions
- WelfordNormalizer: Welford's algorithm with Bessel's correction — suitable for stationary distributions
MLP Learner
Multi-layer perceptron for nonlinear function approximation (Elsayed et al., 2024):
from alberta_framework import MLPLearner, ObGDBounding, RandomWalkStream, run_mlp_learning_loop
import jax.random as jr
stream = RandomWalkStream(feature_dim=10)
learner = MLPLearner(
hidden_sizes=(128, 128),
step_size=1.0,
bounder=ObGDBounding(kappa=2.0),
sparsity=0.9,
)
state, metrics = run_mlp_learning_loop(learner, stream, num_steps=10000, key=jr.key(42))
Streams
Non-stationary experience generators implementing the ScanStream protocol:
RandomWalkStream: Gradual target driftAbruptChangeStream: Sudden target switchesPeriodicChangeStream: Sinusoidal oscillationDynamicScaleShiftStream: Time-varying feature scalesScaleDriftStream: Continuous feature scale drift
TD Learning
For temporal-difference learning with value function approximation:
from alberta_framework import TDLinearLearner, TDIDBD, run_td_learning_loop
learner = TDLinearLearner(optimizer=TDIDBD(trace_decay=0.9))
state, metrics = run_td_learning_loop(learner, td_stream, num_steps=10000, key=jr.key(42))
Gymnasium Integration
from alberta_framework.streams.gymnasium import collect_trajectory, learn_from_trajectory, PredictionMode
import gymnasium as gym
env = gym.make("CartPole-v1")
observations, targets = collect_trajectory(env, policy, num_steps=10000, mode=PredictionMode.REWARD)
state, metrics = learn_from_trajectory(learner, observations, targets)
Publication Tools
Multi-seed experiments with statistical analysis and publication-ready outputs:
from alberta_framework.utils import ExperimentConfig, run_multi_seed_experiment, pairwise_comparisons
results = run_multi_seed_experiment(configs, seeds=30, parallel=True)
significance = pairwise_comparisons(results, test="ttest", correction="bonferroni")
Documentation
Full documentation available at j-klawson.github.io/alberta-framework or build locally:
pip install alberta-framework[docs]
mkdocs serve # http://localhost:8000
Contributing
Contributions are welcome, particularly for upcoming roadmap steps. Please ensure tests pass and follow the existing code style.
pytest tests/ -v
Citation
If you use this framework in your research, please cite:
@software{alberta_framework,
title = {Alberta Framework: A JAX Implementation of Alberta Plan components},
author = {Lawson, Keith},
year = {2026},
url = {https://github.com/j-klawson/alberta-framework}
}
Key References
@article{sutton2022alberta,
title = {The Alberta Plan for AI Research},
author = {Sutton, Richard S. and Bowling, Michael and Pilarski, Patrick M.},
year = {2022},
eprint = {2208.11173},
archivePrefix = {arXiv}
}
@inproceedings{sutton1992idbd,
title = {Adapting Bias by Gradient Descent: An Incremental Version of Delta-Bar-Delta},
author = {Sutton, Richard S.},
booktitle = {Proceedings of the AAAI Conference on Artificial Intelligence},
year = {1992}
}
@inproceedings{mahmood2012autostep,
title = {Tuning-free Step-size Adaptation},
author = {Mahmood, A. Rupam and Sutton, Richard S. and Degris, Thomas and Pilarski, Patrick M.},
booktitle = {IEEE International Conference on Acoustics, Speech and Signal Processing},
year = {2012}
}
@inproceedings{kearney2019tidbd,
title = {Learning Feature Relevance Through Step Size Adaptation in Temporal-Difference Learning},
author = {Kearney, Alex and Veeriah, Vivek and Travnik, Jaden and Sutton, Richard S. and Pilarski, Patrick M.},
booktitle = {International Conference on Machine Learning},
year = {2019}
}
@article{elsayed2024streaming,
title = {Streaming Deep Reinforcement Learning Finally Works},
author = {Elsayed, Mohamed and Lan, Gautham and Lim, Shuze and Mahmood, A. Rupam},
journal = {arXiv preprint arXiv:2410.14606},
year = {2024}
}
License
Apache License 2.0
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file alberta_framework-0.7.0.tar.gz.
File metadata
- Download URL: alberta_framework-0.7.0.tar.gz
- Upload date:
- Size: 132.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
51350dc276281d68629b576a0d23a37c08f97b0aa243d84d58710e215175fbb7
|
|
| MD5 |
5379cff147fd98f0cc45b6da87b8c3da
|
|
| BLAKE2b-256 |
2d8f6de67c778aaf167782e7e42a51a9e07687c91476f5faf69b62d9f9b8f0cf
|
Provenance
The following attestation bundles were made for alberta_framework-0.7.0.tar.gz:
Publisher:
publish.yml on j-klawson/alberta-framework
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
alberta_framework-0.7.0.tar.gz -
Subject digest:
51350dc276281d68629b576a0d23a37c08f97b0aa243d84d58710e215175fbb7 - Sigstore transparency entry: 927238168
- Sigstore integration time:
-
Permalink:
j-klawson/alberta-framework@5b01470edb7bd9e8382b270100f56c6bba72e71a -
Branch / Tag:
refs/tags/v0.7.0 - Owner: https://github.com/j-klawson
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@5b01470edb7bd9e8382b270100f56c6bba72e71a -
Trigger Event:
push
-
Statement type:
File details
Details for the file alberta_framework-0.7.0-py3-none-any.whl.
File metadata
- Download URL: alberta_framework-0.7.0-py3-none-any.whl
- Upload date:
- Size: 68.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
39a91164d99652544ff3c427bb4469d78a3eff3bdd41ae7f7ff7a98afedaa3ca
|
|
| MD5 |
576ed1f7ca4bbda8d8f754bc0a6c102b
|
|
| BLAKE2b-256 |
9afd79b83606b14988232ab9e9b475a2fdc57f46cccdf257015b77269d5cab6f
|
Provenance
The following attestation bundles were made for alberta_framework-0.7.0-py3-none-any.whl:
Publisher:
publish.yml on j-klawson/alberta-framework
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
alberta_framework-0.7.0-py3-none-any.whl -
Subject digest:
39a91164d99652544ff3c427bb4469d78a3eff3bdd41ae7f7ff7a98afedaa3ca - Sigstore transparency entry: 927238171
- Sigstore integration time:
-
Permalink:
j-klawson/alberta-framework@5b01470edb7bd9e8382b270100f56c6bba72e71a -
Branch / Tag:
refs/tags/v0.7.0 - Owner: https://github.com/j-klawson
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@5b01470edb7bd9e8382b270100f56c6bba72e71a -
Trigger Event:
push
-
Statement type: