Parallel Reinforcement Learning library
Project description
PRLearn
PRLearn is a Python library designed for Parallel Reinforcement Learning. It leverages multiprocessing to streamline the training of RL agents, making it easier and more efficient to experiment and develop new RL approaches.
Key Features
- Simple and Flexible: Easy-to-use API built on Gymnasium, enabling seamless integration with existing environments.
- No Dependencies: No mandatory dependencies, but optional use of multiprocess for enhanced parallelism.
- Parallel Data Collection and Training: Efficiently collects and processes data using parallel execution, reducing training time.
- Agent Combination: Combines multiple agents to enhance learning outcomes through methods like averaging, boosting performance and stability.
Installation
Install PRLearn using pip:
pip install prlearn
or
pip install prlearn[multiprocess]
Quick Start
Here's a brief example to get you started with PRLearn:
Define Your Agent
from prlearn import Agent, Experience
from typing import Any, Dict, Tuple
class MyAgent(Agent):
def action(self, state: Tuple[Any, Dict[str, Any]]) -> Any:
observation, info = state
# Define action logic
pass
def train(self, experience: Experience):
obs, actions, rewards, terminated, truncated, info = experience.get()
# Define training logic
pass
Using Parallel Training
import gymnasium as gym
from my_best_model import BestAgent
from prlearn import Trainer
from prlearn.collection.agent_combiners import FixedStatAgentCombiner
# Define your environment and agent
env = gym.make("LunarLander-v2")
agent = BestAgent()
# Create and configure the trainer
trainer = Trainer(
agent=agent,
env=env,
n_workers=4,
schedule=[
("train_finish", 1000, "episodes"),
("train_agent", 10, "episodes"),
],
mode="parallel_learning", # optional
sync_mode="sync", # optional
combiner=FixedStatAgentCombiner("mean_reward"), # optional
)
# Run the trainer
agent, result = trainer.run()
- Environment: We use the
LunarLander-v2
environment from Gymnasium. - Agent:
BestAgent
is a custom agent class you should define. - Trainer Configuration: The trainer is configured with 4 parallel workers, a schedule that specifies training completion after 1000 episodes and agent training every 10 episodes.
- Optional parameters include the mode
parallel_learning
, synchronization modesync
, and a combinerFixedStatAgentCombiner
that averages agent rewards.
License
Licensed under the MIT License.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
prlearn-0.0.1.tar.gz
(14.0 kB
view details)
Built Distribution
prlearn-0.0.1-py3-none-any.whl
(17.7 kB
view details)
File details
Details for the file prlearn-0.0.1.tar.gz
.
File metadata
- Download URL: prlearn-0.0.1.tar.gz
- Upload date:
- Size: 14.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.7.1 CPython/3.8.11 Darwin/23.4.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 3c4eff79f41722d0ac6969ed17baa2e681498a1dee80aeafd5092b54e426f929 |
|
MD5 | 5add5334f76436b099c712b7cd838e25 |
|
BLAKE2b-256 | b5eacdd5cbf0d94d635166f9d192b7fc8793ae44c68d828b416422b9550dd443 |
Provenance
File details
Details for the file prlearn-0.0.1-py3-none-any.whl
.
File metadata
- Download URL: prlearn-0.0.1-py3-none-any.whl
- Upload date:
- Size: 17.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.7.1 CPython/3.8.11 Darwin/23.4.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 73e5dfeed508cae639ba709dd97e984a7272e16d5fb24990a3c94716650fd786 |
|
MD5 | 4c9de7967333a3fab37e20d1869bab6d |
|
BLAKE2b-256 | 110720db68fca9cceec6c7cef8fc456f325cd0ca9764566c22660da347d4eb2c |