Skip to main content

Monarch: Single controller library

Project description

Monarch 🦋

Monarch is a distributed programming framework for PyTorch based on scalable actor messaging. It provides:

  1. Remote actors with scalable messaging: Actors are grouped into collections called meshes and messages can be broadcast to all members.
  2. Fault tolerance through supervision trees: Actors and processes form a tree and failures propagate up the tree, providing good default error behavior and enabling fine-grained fault recovery.
  3. Point-to-point RDMA transfers: cheap registration of any GPU or CPU memory in a process, with the one-sided transfers based on libibverbs
  4. Distributed tensors: actors can work with tensor objects sharded across processes

Monarch code imperatively describes how to create processes and actors using a simple python API:

from monarch.actor import Actor, endpoint, this_host

# spawn 8 trainer processes one for each gpu
training_procs = this_host().spawn_procs({"gpus": 8})


# define the actor to run on each process
class Trainer(Actor):
    @endpoint
    def train(self, step: int): ...


# create the trainers
trainers = training_procs.spawn("trainers", Trainer)

# tell all the trainers to take a step
fut = trainers.train.call(step=0)

# wait for all trainers to complete
fut.get()

The introduction to monarch concepts provides an introduction to using these features.

⚠️ Early Development Warning Monarch is currently in an experimental stage. You should expect bugs, incomplete features, and APIs that may change in future versions. The project welcomes bugfixes, but to make sure things are well coordinated you should discuss any significant change before starting the work. It's recommended that you signal your intention to contribute in the issue tracker, either by filing a new issue or by claiming an existing one.

📖 Documentation

View Monarch's hosted documentation at this link.

Installation

Installing from Pre-built Wheels

Monarch provides pre-built wheels that work regardless of what version of PyTorch you have installed:

Stable

pip install torchmonarch

Nightly

pip install --pre torchmonarch

Or install a specific nightly version:

pip install torchmonarch==0.3.0.dev20260106

Build and Install from Source

Note: Building from source requires additional system dependencies. These are needed at build time only, not at runtime.

Monarch uses uv for fast, reliable Python package management. If you don't have uv installed:

# Install uv
curl -LsSf https://astral.sh/uv/install.sh | sh
# Or on macOS
brew install uv

Configuring PyTorch Index: By default, Monarch builds with PyTorch from the pytorch-cu128 index (CUDA 12.8). To use a different CUDA version:

  • Edit [tool.uv.sources] in pyproject.toml to point to a different index (e.g., pytorch-cu126, pytorch-cu130, or pytorch-cpu)
  • Or use --extra-index-url when running uv:
    uv sync --extra-index-url https://download.pytorch.org/whl/cu126
    

Understanding Tensor Engine

Monarch includes distributed tensor and RDMA APIs. Since these are hardware-specific, it can be useful to develop with a lighter-weight version of Monarch (actors only) by setting USE_TENSOR_ENGINE=0.

By default, Monarch builds with tensor_engine enabled. To build without it:

USE_TENSOR_ENGINE=0 uv sync

Note: Building without tensor_engine means you won't have access to the distributed tensor or RDMA APIs. Torch is required to use tensor_engine, and the latest stable torch is ABI compatible with the latest versioned torchmonarch

Build Dependencies by Platform

On Fedora distributions
# Install nightly rust toolchain
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh
rustup toolchain install nightly
rustup default nightly

# Install non-python dependencies
sudo dnf install -y cmake ninja-build protobuf-compiler libunwind

# Install the correct cuda and cuda-toolkit versions for your machine
sudo dnf install cuda-toolkit-12-8 cuda-12-8

# Install clang-devel, nccl-devel, and libstdc++-static
sudo dnf install clang-devel libnccl-devel libstdc++-static

# Install RDMA libraries (needed for tensor_engine builds)
sudo dnf install -y libibverbs rdma-core libmlx5 libibverbs-devel rdma-core-devel

# Clone and sync dependencies
git clone https://github.com/meta-pytorch/monarch.git
cd monarch

# Install in development mode with all dependencies
uv sync

# Or install without tensor_engine
USE_TENSOR_ENGINE=0 uv sync

# Verify installation
uv run python -c "from monarch import actor; print('Monarch installed successfully')"

# Rebuild (e.g., after changing Rust code)
USE_TENSOR_ENGINE=0 uv pip install -e .
On Ubuntu distributions
# Install nightly rust toolchain
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh
source $HOME/.cargo/env
rustup toolchain install nightly
rustup default nightly

# Install Ubuntu-specific system dependencies
sudo apt install -y cmake ninja-build protobuf-compiler libunwind-dev clang

# Set clang as the default C/C++ compiler
export CC=clang
export CXX=clang++

# Install the correct cuda and cuda-toolkit versions for your machine
sudo apt install -y cuda-toolkit-12-8 cuda-12-8

# Install RDMA libraries (needed for tensor_engine builds)
sudo apt install -y rdma-core libibverbs1 libmlx5-1 libibverbs-dev

# Clone and sync dependencies
git clone https://github.com/meta-pytorch/monarch.git
cd monarch

# Install in development mode with all dependencies
uv sync

# Or install without tensor_engine (CPU-only)
USE_TENSOR_ENGINE=0 uv sync

# Verify installation
uv run python -c "from monarch import actor; print('Monarch installed successfully')"

# Rebuild (e.g., after changing Rust code)
USE_TENSOR_ENGINE=0 uv pip install -e .
On non-CUDA machines

You can also build Monarch on non-CUDA machines (e.g., macOS laptops) for CPU-only usage.

Note that this does not support tensor_engine, which requires CUDA and RDMA libraries.

# Install nightly rust toolchain
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh
rustup toolchain install nightly
rustup default nightly

# Clone and sync dependencies (without tensor_engine)
git clone https://github.com/meta-pytorch/monarch.git
cd monarch

# Install without tensor engine (CPU-only)
USE_TENSOR_ENGINE=0 uv sync

# Verify installation
uv run python -c "from monarch import actor; print('Monarch installed successfully')"

Alternative: Using pip

If you prefer to use pip instead of uv:

# After installing system dependencies (see above)

# Install build dependencies

# Build and install Monarch
pip install .

# Or for development
pip install -e .

# Without tensor_engine
USE_TENSOR_ENGINE=0 pip install -e .

Running examples

Check out the examples/ directory for demonstrations of how to use Monarch's APIs.

We'll be adding more examples as we stabilize and polish functionality!

Running tests

We have both Rust and Python unit tests. Rust tests are run with cargo-nextest and Python tests are run with pytest.

Rust tests

Important: Monarch's Rust code uses PyO3 to interface with Python, which means the Rust binaries need to link against Python libraries. Before running Rust tests, you need to have a Python environment activated (conda, venv, or uv):

# If using uv (recommended)
uv sync  # This creates and activates a virtual environment
uv run cargo nextest run  # Run tests within the uv environment

# Or if using conda
conda activate monarchenv
cargo nextest run

# Or if using venv
source .venv/bin/activate
cargo nextest run

Without an active Python environment, you'll get Python linking errors like:

error: could not find native static library `python3.12`, perhaps an -L flag is missing?

Installing cargo-nextest:

# We use cargo-nextest to run our tests, as they provide strong process isolation
# between every test.
# Here we install it from source, but you can instead use a pre-built binary described
# here: https://nexte.st/docs/installation/pre-built-binaries/
cargo install cargo-nextest --locked

cargo-nextest supports all of the filtering flags of "cargo test".

Python tests

# Install test dependencies (if not already installed via uv sync)
uv sync --extra test

# Run unit tests with uv
uv run pytest python/tests/ -v -m "not oss_skip"

# Or if using pip
pip install -e '.[test]'
pytest python/tests/ -v -m "not oss_skip"

Disabling flaky CI tests

If a test is consistently failing in OSS CI and needs to be temporarily disabled without a code change, open a GitHub issue on this repo with a title of the form:

DISABLED <test-name>

At the start of each CI run, scripts/fetch_disabled_tests.py fetches all open issues whose titles start with DISABLED and skips the named tests. Closing the issue re-enables the test on the next run.

Naming format:

  • Rust (cargo nextest): use the test name exactly as it appears in nextest output: <binary> <module::path::test_fn>, e.g. DISABLED hyperactor proc::tests::test_child_lifecycle
  • Python (pytest): use the test function name, e.g. DISABLED test_my_function

License

Monarch is BSD-3 licensed, as found in the LICENSE file.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distributions

If you're not sure about the file name format, learn more about wheel file names.

File details

Details for the file torchmonarch-0.4.0.dev20260324-cp313-cp313-manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for torchmonarch-0.4.0.dev20260324-cp313-cp313-manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 f991521542785c23cc321d5849ec03b069541623ed2249853959b5cc97cacba3
MD5 128f68da073641862fc48e69f9da5735
BLAKE2b-256 f35a1a7e1afa1bf3fe45d8b4a1a175e15601b8df9970fe148208f2b7bf28aede

See more details on using hashes here.

Provenance

The following attestation bundles were made for torchmonarch-0.4.0.dev20260324-cp313-cp313-manylinux2014_x86_64.whl:

Publisher: wheels.yml on meta-pytorch/monarch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file torchmonarch-0.4.0.dev20260324-cp313-cp313-manylinux2014_aarch64.whl.

File metadata

File hashes

Hashes for torchmonarch-0.4.0.dev20260324-cp313-cp313-manylinux2014_aarch64.whl
Algorithm Hash digest
SHA256 16c3876f77a8b89808bd0a3a7d4f8721a47021799dd44d03403b1ec15f2ecef8
MD5 24317daf9f0a9e1dd92506ff6ca79a4c
BLAKE2b-256 030221def4bad63f82edce21d0d3c40c9e3a3fa4a24eb25ea1d1d76e4c6c02e2

See more details on using hashes here.

Provenance

The following attestation bundles were made for torchmonarch-0.4.0.dev20260324-cp313-cp313-manylinux2014_aarch64.whl:

Publisher: wheels.yml on meta-pytorch/monarch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file torchmonarch-0.4.0.dev20260324-cp312-cp312-manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for torchmonarch-0.4.0.dev20260324-cp312-cp312-manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 68a0d453d5e7545c3d726870149c5bc15d90855ad0751a037536d4f9f78ef03f
MD5 3f60c3b0b1fb469a66204276ee3d0976
BLAKE2b-256 e796ec2555ed3f33339e9acb6aedbf612effb44fa34dc6505b3d0d860bb445d3

See more details on using hashes here.

Provenance

The following attestation bundles were made for torchmonarch-0.4.0.dev20260324-cp312-cp312-manylinux2014_x86_64.whl:

Publisher: wheels.yml on meta-pytorch/monarch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file torchmonarch-0.4.0.dev20260324-cp312-cp312-manylinux2014_aarch64.whl.

File metadata

File hashes

Hashes for torchmonarch-0.4.0.dev20260324-cp312-cp312-manylinux2014_aarch64.whl
Algorithm Hash digest
SHA256 251112e73adc69c0cb1cd6f871f1fb2ed48e228aa2491c2373c7bba6cb50ede1
MD5 439256402ca313302e1235c86d8a6df2
BLAKE2b-256 d849b9837d306b54498b8279e3d0a698208977b9b98f3cad076c28a48f2119ca

See more details on using hashes here.

Provenance

The following attestation bundles were made for torchmonarch-0.4.0.dev20260324-cp312-cp312-manylinux2014_aarch64.whl:

Publisher: wheels.yml on meta-pytorch/monarch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file torchmonarch-0.4.0.dev20260324-cp311-cp311-manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for torchmonarch-0.4.0.dev20260324-cp311-cp311-manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 3d68c16458e8cd5ed6ed7d53fc18b67c3151acce42bb9ca2add1ceccb5ff7311
MD5 c819ee57e14bbb203ea03495d4fc779a
BLAKE2b-256 46391be498e67d0d36a430fd639eceda4c9e08ad081ef69ba6a78ff32ad6a456

See more details on using hashes here.

Provenance

The following attestation bundles were made for torchmonarch-0.4.0.dev20260324-cp311-cp311-manylinux2014_x86_64.whl:

Publisher: wheels.yml on meta-pytorch/monarch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file torchmonarch-0.4.0.dev20260324-cp311-cp311-manylinux2014_aarch64.whl.

File metadata

File hashes

Hashes for torchmonarch-0.4.0.dev20260324-cp311-cp311-manylinux2014_aarch64.whl
Algorithm Hash digest
SHA256 d1e5dc617ab7229892d8ace02ca4977d896f6ea52fedaf8cca94e14afde237fa
MD5 c905f48c08108a4f26d59f0cfc5a9996
BLAKE2b-256 9ef7b41ac9fd5724fdec6616be13b49caa32135ff9c30210bf81647c259de967

See more details on using hashes here.

Provenance

The following attestation bundles were made for torchmonarch-0.4.0.dev20260324-cp311-cp311-manylinux2014_aarch64.whl:

Publisher: wheels.yml on meta-pytorch/monarch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file torchmonarch-0.4.0.dev20260324-cp310-cp310-manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for torchmonarch-0.4.0.dev20260324-cp310-cp310-manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 7abadb00e45b66a454669ce0b153eb38b23c3475ddc5fb96e920320fdf83e0e2
MD5 f48a66c3967148dbbac9361350b18bbe
BLAKE2b-256 cbd943221825c045862dab817fd8a628889de9e03d5662b0235fa6a6316ef8c6

See more details on using hashes here.

Provenance

The following attestation bundles were made for torchmonarch-0.4.0.dev20260324-cp310-cp310-manylinux2014_x86_64.whl:

Publisher: wheels.yml on meta-pytorch/monarch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file torchmonarch-0.4.0.dev20260324-cp310-cp310-manylinux2014_aarch64.whl.

File metadata

File hashes

Hashes for torchmonarch-0.4.0.dev20260324-cp310-cp310-manylinux2014_aarch64.whl
Algorithm Hash digest
SHA256 239bac6f27d3aee1fd2712aee8077e90dbc628e89c8bd868b79aa0093816d25a
MD5 e4bad7ec20fb4fa4ecfc87fb61533758
BLAKE2b-256 642e4b71d8f1d9672cfd2daebe5a56dd17f34801b2f849be29ee7b3e1f4701c6

See more details on using hashes here.

Provenance

The following attestation bundles were made for torchmonarch-0.4.0.dev20260324-cp310-cp310-manylinux2014_aarch64.whl:

Publisher: wheels.yml on meta-pytorch/monarch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page