A Python library for reinforcement learning algorithms
Project description
🎯 NeatRL
A clean, modern Python library for reinforcement learning algorithms
NeatRL provides high-quality implementations of popular RL algorithms with a focus on simplicity, performance, and ease of use. Built with PyTorch and designed for both research and production use.
✨ Features
- 📊 Experiment Tracking: Built-in support for Weights & Biases logging
- 🎮 Gymnasium Compatible: Works with Gymnasium environments and adding many more!
- 🎯 Atari Support: Full support for Atari games with automatic CNN architectures
- ⚡ Parallel Training: Vectorized environments for faster data collection
- 🔧 Easy to Extend: Modular design for adding new algorithms
- 📈 State-of-the-Art: Implements modern RL techniques and best practices
- 🎥 Video Recording: Automatic video capture and WandB integration
- 📉 Advanced Logging: Per-layer gradient monitoring and comprehensive metrics
🏗️ Supported Algorithms
Current Implementations
-
DQN (Deep Q-Network) - Classic value-based RL algorithm
- Support for discrete action spaces
- Experience replay and target networks
- Atari preprocessing and frame stacking
-
Dueling DQN - Enhanced DQN with separate value and advantage streams
- Improved learning stability
- Better performance on complex environments
-
REINFORCE - Policy gradient method for discrete and continuous action spaces
- Atari game support with automatic CNN architecture
- Parallel environment training (
n_envssupport) - Continuous action space support
- Episode-based Monte Carlo returns
- Variance reduction through baseline subtraction
-
DDPG (Deep Deterministic Policy Gradient) - Actor-critic method for continuous action spaces
- Deterministic policy gradient for continuous control
- Experience replay and target networks
- Ornstein-Uhlenbeck noise for exploration
- Support for exact continuous action spaces
-
A2C (Advantage Actor-Critic) - Synchronous actor-critic algorithm
- Synchronous version of A3C for stable training
- Advantage function for reduced variance
- Support for both discrete and continuous action spaces
- Parallel environment training with vectorized environments
- Monte Carlo returns for value estimation
-
TD3 (Twin Delayed DDPG) - Advanced actor-critic method for continuous control
- Twin Q-networks to reduce overestimation bias
- Delayed policy updates for improved stability
- Target policy smoothing with noise
- Experience replay and target networks
- CNN support for image-based environments
-
SAC (Soft Actor-Critic) - Maximum entropy reinforcement learning
- Stochastic Gaussian policies with entropy regularization
- Twin Q-networks for stable learning
- Automatic entropy tuning (alpha parameter)
- Balances exploration and exploitation
- CNN support for complex environments
-
PPO (Proximal Policy Optimization) - State-of-the-art policy gradient method with GAE
- Full PPO implementation with Generalized Advantage Estimation (GAE)
- Support for both discrete and continuous action spaces
- Atari game support with automatic CNN architecture
- Clipped surrogate objective for stable policy updates
- Value function clipping and entropy regularization
- Vectorized environments for parallel training
-
PPO-RND (Proximal Policy Optimization with Random Network Distillation) - State-of-the-art exploration method
- Intrinsic motivation through novelty detection
- Combined extrinsic and intrinsic rewards for better exploration
- Support for both discrete and continuous action spaces
- PPO with clipped surrogate objective
- Vectorized environments for parallel training
- Intrinsic reward normalization and advantage calculation
-
More algorithms coming soon...
📦 Installation
python -m venv neatrl-env
source neatrl-env/bin/activate
pip install neatrl
# Install extras based on environments you want to use
pip install neatrl[atari] # For CarRacing
pip install neatrl[box2d] # For BipedalWalker
pip install neatrl[classic] # For Pendulum
pip install neatrl[mujoco] # For HalfCheetah
🚀 Quick Start
Train DQN on CartPole
from neatrl import train_dqn
model = train_dqn(
env_id="CartPole-v1",
total_timesteps=10000,
seed=42
)
Train PPO on Classic Control
from neatrl import train_ppo
model = train_ppo(
env_id="CartPole-v1",
total_timesteps=50000,
n_envs=4, # Parallel environments
GAE=0.95, # Generalized Advantage Estimation lambda
clip_value=0.2, # PPO clipping parameter
use_wandb=True, # Track with WandB
seed=42
)
Train SAC on Continuous Control
from neatrl import train_sac
model = train_sac(
env_id="Pendulum-v1",
total_timesteps=50000,
alpha=0.2, # Entropy regularization coefficient
autotune_alpha=True, # Automatically tune alpha
use_wandb=True, # Track with WandB
seed=42
)
Train SAC on Atari
from neatrl import train_sac_cnn
model = train_sac_cnn(
env_id="BreakoutNoFrameskip-v4",
total_timesteps=100000,
alpha=0.2,
autotune_alpha=True,
atari_wrapper=True, # Automatic Atari preprocessing
use_wandb=True, # Track with WandB
seed=42
)
📚 Documentation
The docs include:
- Detailed usage examples
- Hyperparameter tuning guides
- Environment compatibility
- Experiment tracking setup
- Troubleshooting tips
🤝 Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
Development Setup
git clone https://github.com/YuvrajSingh-mist/NeatRL.git
cd NeatRL
pip install -e .[dev]
For the complete changelog, see CHANGELOG.md.
📄 License
This project is licensed under the MIT License - see the LICENSE file for details.
Made with ❤️ for the RL community
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file neatrl-1.0.0.tar.gz.
File metadata
- Download URL: neatrl-1.0.0.tar.gz
- Upload date:
- Size: 63.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
b8cdd1ef95c5b00e56d66688989429b5d9d16be0c425bf3137a78814c29f10a1
|
|
| MD5 |
806f5b4657112558e64de2f3c97ea6a7
|
|
| BLAKE2b-256 |
be37c2b1dfdd9cc55c4fa21a9ea336e20da311b299e8b062d9a3a7c4e917da0e
|
File details
Details for the file neatrl-1.0.0-py3-none-any.whl.
File metadata
- Download URL: neatrl-1.0.0-py3-none-any.whl
- Upload date:
- Size: 73.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
36e13d47172433e9fd32f20d0dfff6716b5e50f57e3f84e10c62f317b07e861e
|
|
| MD5 |
e1546f72e71c7b1818922fb97549e158
|
|
| BLAKE2b-256 |
c2e86b8bf0d4a5aadafa0c0a71d633dbd6e66658407a8596a648aee7ae780cba
|