Regelum is a flexibly configurable framework for agent-environment simulation with a menu of predictive and reinforcement learning pipelines.
Project description
About
Regelum-control
stands as a framework designed to address optimal control and reinforcement learning (RL) tasks within continuous-time dynamical systems. It is made for researchers and engineers in reinforcement learning and control theory.
A detailed documentation is available here.
Explore regelum-playground repo with ready-to-use examples.
Features
-
Run pre-configured regelum algorithms with ease. Regelum offers a set of implemented, ready-to-use algorithms in the domain of RL and Optimal Control. It provides flexibility through multiple optimization backends, including CasADi and PyTorch, to accommodate various computational needs.
-
Stabilize your dynamical system with Regelum. Regelum stands as a framework designed to address optimal control and reinforcement learning (RL) tasks within continuous-time dynamical systems. It comes equipped with an array of default systems, alongside a detailed tutorial that provides clear instructions for users to instantiate their own environments.
-
Manage your experiment data. Regelum seamlessly captures every detail of your experiment with little to no configuration required. From parameters to performance metrics, every datum is recorded. Through integration with MLflow, Regelum streamlines tracking, comparison and real-time monitoring of metrics.
-
Reproduce your experiments with ease. Commit hashes and diffs for every experiment are also stored in Regelum, offering the ability to reproduce your experiments at any time with simple terminal commands.
-
Configure your experiments efficiently. Our Hydra fork within Regelum introduces enhanced functionaly, making the configuration of your RL and Optimal Control tasks more accessible and user-friendly.
-
Fine-tune your models to perfection and achieve peak performance with minimal effort. By integrating with Hydra, regelum inherently adopts Hydra's powerful hyperparameter tuning capability.
Install regelum-control with pip
pip install regelum-control
Developer setup
- Clone the repository.
- Run:
pip install -e . bash scripts/developer-setup.sh
- Check
requirements-dev.txt
in the root of the repo for additional details.
Licence
This project is licensed under the terms of the MIT license.
Bibtex reference
@misc{regelum2024,
author = {Pavel Osinenko, Grigory Yaremenko, Georgiy Malaniya, Anton Bolychev},
title = {Regelum: a framework for simulation, control and reinforcement learning},
howpublished = {\url{https://github.com/osinekop/regelum-control}},
year = {2024},
note = {Licensed under the MIT License}
}
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file regelum_control-0.3.3.tar.gz
.
File metadata
- Download URL: regelum_control-0.3.3.tar.gz
- Upload date:
- Size: 135.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.4 CPython/3.10.14 Linux/6.8.0-45-generic
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | e85fe66c1fa7736e296b93ed094701041b47baead9a7fc1ea0be8cdb83458ffc |
|
MD5 | 69c40c89eddb7399852ea4b3f374dd03 |
|
BLAKE2b-256 | 2c1a4b15742eea6f7e4aec69d97df0a692b77c0865b2552819dc30630f52b878 |
File details
Details for the file regelum_control-0.3.3-py3-none-any.whl
.
File metadata
- Download URL: regelum_control-0.3.3-py3-none-any.whl
- Upload date:
- Size: 148.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.4 CPython/3.10.14 Linux/6.8.0-45-generic
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | bebe93fb92acade966fea94da94c27f6c16469a36697b2887740b32a36109048 |
|
MD5 | 7419aef47b3318a026a71b37b7b8ae67 |
|
BLAKE2b-256 | b0875268ba5a4a34d6d16ce749d07511244646fcd98151605967372e4c480d4d |