Skip to main content

A strongly typed Multi-Agent Reinforcement Learning framework

Project description

marlenv - A unified framework for muti-agent reinforcement learning

Documentation: https://yamoling.github.io/multi-agent-rlenv

The objective of marlenv is to provide a common (typed) interface for many different reinforcement learning environments.

As such, marlenv provides high level abstractions of RL concepts such as Observations or Transitions that are commonly represented as mere (confusing) lists or tuples.

Installation

Install with you preferred package manager (uv, pip, poetry, ...):

$ pip install marlenv[all] # Enable all features
$ pip install marlenv      # Basic installation

There are multiple optional dependencies if you want to support specific libraries and environments. Available options are:

  • smac for StarCraft II environments
  • gym for OpenAI Gym environments
  • pettingzoo for PettingZoo environments
  • overcooked for Overcooked environments

Install them with:

$ pip install marlenv[smac] # Install SMAC
$ pip install marlenv[gym,smac]  # Install Gym & smac support

Using the marlenv environment catalog

Some environments are registered in the marlenv and can be easily instantiated via its catalog.

from marlenv import catalog

env1 = catalog.Overcooked.from_layout("scenario4")
env2 = catalog.LLE.level(6)
env3 = catalog.DeepSea(mex_depth=5)

Note that using the catalog requires the corresponding environment package to be installed. For instance you need to install the laser-learning-environment package to use catalog.LLE, which can be done by using the corresponding feature when at installation as shown below.

pip install multi-agent-rlenv[lle]

Using marlenv with existing libraries

marlenv provides adapters from most popular libraries to unify them under a single interface. Namely, marlenv supports smac, gymnasium and pettingzoo.

import marlenv

# You can instanciate gymnasium environments directly via their registry ID
gym_env = marlenv.make("CartPole-v1", seed=25)

# You can seemlessly instanciate a SMAC environment and directly pass your required arguments
from marlenv.adapters import SMAC
smac_env = SMAC("3m", debug=True, difficulty="9")

# pettingzoo is also supported
from pettingzoo.sisl import pursuit_v4
from marlenv.adapters import PettingZoo
pz_env = PettingZoo(pursuit_v4.parallel_env())

Designing custom environments

You can create your own custom environment by inheriting from the RLEnv class. The below example illustrates a gridworld with a discrete action space. Note that other methods such as step or render must also be implemented.

import numpy as np
from marlenv import RLEnv, DiscreteActionSpace, Observation

N_AGENTS = 3
N_ACTIONS = 5

class CustomEnv(MARLEnv[DiscreteActionSpace]):
    def __init__(self, width: int, height: int):
        super().__init__(
            action_space=DiscreteActionSpace(N_AGENTS, N_ACTIONS),
            observation_shape=(height, width),
            state_shape=(1,),
        )
        self.time = 0

    def reset(self) -> Observation:
        self.time = 0
        ...
        return obs

    def get_state(self):
        return np.array([self.time])

Useful wrappers

marlenv comes with multiple common environment wrappers, check the documentation for a complete list. The preferred way of using the wrappers is through a marlenv.Builder. The below example shows how to add a time limit (in number of steps) and an agent id to the observations of a SMAC environment.

from marlenv import Builder
from marlenv.adapters import SMAC

env = Builder(SMAC("3m")).agent_id().time_limit(20).build()
print(env.extras_shape) # -> (4, ) because there are 3 agents and the time counter

Related projects

Project details


Release history Release notifications | RSS feed

This version

3.7.0

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

multi_agent_rlenv-3.7.0.tar.gz (43.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

multi_agent_rlenv-3.7.0-py3-none-any.whl (47.2 kB view details)

Uploaded Python 3

File details

Details for the file multi_agent_rlenv-3.7.0.tar.gz.

File metadata

  • Download URL: multi_agent_rlenv-3.7.0.tar.gz
  • Upload date:
  • Size: 43.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for multi_agent_rlenv-3.7.0.tar.gz
Algorithm Hash digest
SHA256 6cbb5f0dca4230fb13756fea2d0979322b500bcd5be671346830eedb35824b30
MD5 7ba436bac6c7cb7d6498c7dc7fc7de15
BLAKE2b-256 248080e61d40af209361f959d9d75f51ec4382a48d5d63806f66eb388b8c7a6e

See more details on using hashes here.

Provenance

The following attestation bundles were made for multi_agent_rlenv-3.7.0.tar.gz:

Publisher: ci.yaml on yamoling/multi-agent-rlenv

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file multi_agent_rlenv-3.7.0-py3-none-any.whl.

File metadata

File hashes

Hashes for multi_agent_rlenv-3.7.0-py3-none-any.whl
Algorithm Hash digest
SHA256 69a9de74515c60970a6f11b5a23f6d82af706c5bd891406e0cedc7cc73341e6d
MD5 a085acd90f28f78749b9e93f9c1fcbb3
BLAKE2b-256 c7029c050389267a99f4b841baf21548d52c38d2dd04b57ece98e8426b7ce194

See more details on using hashes here.

Provenance

The following attestation bundles were made for multi_agent_rlenv-3.7.0-py3-none-any.whl:

Publisher: ci.yaml on yamoling/multi-agent-rlenv

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page