Production-ready runtime for building and orchestrating intelligent multi-agent AI systems
Project description
MUXI Runtime
The execution engine for AI agent formations.
For most users: Install MUXI CLI for the complete experience. This repo is for contributors and developers embedding the runtime directly.
[!IMPORTANT]
MUXI Ecosystem
This repository is part of the larger MUXI ecosystem.
๐ Complete architectural overview: See muxi/ARCHITECTURE.md - explains how core repositories fit together, dependencies, status, and roadmap.
What is MUXI Runtime?
MUXI Runtime transforms declarative YAML configurations into running AI systems. It's the core engine that powers the MUXI Server.
Core responsibilities:
- Formation execution - Loads and runs agent configurations from YAML
- Overlord orchestration - Routes requests, manages clarifications, coordinates workflows
- Memory systems - Three-tier memory (buffer, persistent, vector)
- Tool integration - MCP protocol support for external tools
- Multi-tenant isolation - User and session management
Architecture
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ MUXI Server - Formation lifecycle management โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ
โผ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ MUXI Runtime โโโ THIS REPO โ
โ โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ โ
โ โ Formation Engine (YAML loader & validator) โ โ
โ โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค โ
โ โ Overlord โ Agents โ Workflow โ Background โ โ
โ โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค โ
โ โ Memory โ MCP โ A2A โ LLM โ Observability โ โ
โ โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ
โผ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ External Services (LLM APIs, MCP Servers, DBs) โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
Package Structure
The runtime uses src/muxi/runtime/ to share the muxi namespace with the Python SDK:
src/muxi/runtime/
โโโ formation/ # Formation engine
โ โโโ overlord/ # Central orchestration
โ โโโ agents/ # Agent implementations
โ โโโ workflow/ # Task decomposition, SOPs
โ โโโ server/ # Formation API (FastAPI)
โ โโโ background/ # Webhooks, scheduling, async
โโโ services/ # Runtime services
โ โโโ memory/ # Memory systems
โ โโโ mcp/ # MCP client
โ โโโ a2a/ # Agent-to-agent
โ โโโ llm/ # LLM abstraction
โโโ datatypes/ # Type definitions
Quick Start
Using with MUXI Server (recommended)
# Install MUXI CLI
curl -fsSL https://muxi.ai/install | sh
# Create and run a formation
muxi new my-assistant
cd my-assistant
muxi dev
Embedding directly
# Base install (covers almost all use cases)
pip install muxi-runtime
# PyTorch variant โ for local embedding models without an ONNX export
pip install 'muxi-runtime[pytorch]'
# CUDA variant โ GPU-accelerated ONNX/PyTorch models + FAISS-GPU (NVIDIA only)
# Must uninstall CPU-only packages first to avoid conflicts:
pip uninstall -y faiss-cpu faissx onnxruntime
pip install 'muxi-runtime[cuda]'
from muxi.runtime import Formation
import asyncio
async def main():
formation = Formation()
await formation.load("formation.afs")
overlord = await formation.start_overlord()
response = await overlord.chat(
"Hello!",
user_id="user123"
)
print(response)
asyncio.run(main())
Docker Images
MUXI Runtime ships three image variants. All support linux/amd64 and linux/arm64 (except CUDA).
Most users should use the base variant. The PyTorch and CUDA variants exist for specific embedding workloads described below.
| Variant | SIF size | Description | Status |
|---|---|---|---|
default (base) |
~600 MB | Lean runtime โ covers the vast majority of use cases | Stable |
pytorch |
larger | Adds CPU-only PyTorch for local embedding models that lack ONNX exports | Stable |
cuda |
largest | GPU-accelerated: ONNX and PyTorch local models + FAISS-GPU for faster vector ops | Experimental |
When to use each variant:
defaultโ the right choice for almost everyone. Uses ONNX-based local embedding models (fast, lightweight) and CPU FAISS.pytorchโ only needed when you want to run a local embedding model that does not have an ONNX export and therefore requires the full PyTorch runtime.cudaโ recommended for production workloads running on servers with NVIDIA GPUs. Supports both ONNX and PyTorch local models, and ships with FAISS-GPU for significantly faster vector similarity operations.
# Build the default (base) variant
./scripts/build/runtime.sh
# Build the PyTorch variant (requires default built first)
./scripts/build/runtime.sh --variant pytorch
# Build the CUDA variant (experimental, linux/amd64 + NVIDIA tooling required)
./scripts/build/runtime.sh --variant cuda
# Cross-compile for a specific platform
./scripts/build/runtime.sh --platform linux/amd64 --variant pytorch
SIF (Singularity/Apptainer)
Each variant can be converted to a .sif artifact for use with MUXI Server:
./scripts/build/sif.sh # default
./scripts/build/sif.sh --variant pytorch
./scripts/build/sif.sh --variant cuda # experimental
./scripts/build/sif.sh --arch amd64 # force architecture
Note: On macOS and Windows the correct SIF architecture is always
linux-amd64, regardless of host CPU.linux-arm64SIFs only apply on native arm64 Linux hosts (e.g. AWS Graviton).CUDA variant is experimental. It has not been end-to-end validated against live GPUs in CI and only builds on
linux/amd64hosts with NVIDIA tooling.
Development
git clone https://github.com/muxi-ai/runtime
cd runtime
pip install -e ".[dev]"
# Unit and integration tests
pytest tests/unit -v
pytest tests/integration -v
# E2E tests (standalone scripts, not pytest)
cd e2e && python run_all_tests.py # full suite
cd e2e && python run_random_tests.py 10 # random sample
cd e2e/tests/<area> && python test_<name>.py # single test
See contributing/README.md for contributor documentation.
Related Repositories
| Repo | Description |
|---|---|
| muxi-ai/muxi | Main repo with architecture docs |
| muxi-ai/server | Go server that hosts this runtime |
| muxi-ai/cli | Command-line tool |
| muxi-ai/sdks | Python, TypeScript, Go SDKs |
| muxi-ai/schemas | API schemas |
Documentation
- User docs: docs.muxi.ai
- Contributor docs: contributing/README.md
- Formation spec: agentformation.org
License
Elastic License 2.0 - Free to use, modify, and embed in products. Cannot be offered as a hosted service.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file muxi_runtime-0.20260424.0.tar.gz.
File metadata
- Download URL: muxi_runtime-0.20260424.0.tar.gz
- Upload date:
- Size: 1.5 MB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.13
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
9e75ce23e7f165beca3195de487e8c1a880c9aeb53cad230e1173f606cd4f0d6
|
|
| MD5 |
61501d9497fbb0b7bc4e136df4b41008
|
|
| BLAKE2b-256 |
10b7d1434413bb1eaa62d889a6b2ce51429c9e5363f6221ff479b06c4bd9ccd9
|
Provenance
The following attestation bundles were made for muxi_runtime-0.20260424.0.tar.gz:
Publisher:
release.yml on muxi-ai/runtime
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
muxi_runtime-0.20260424.0.tar.gz -
Subject digest:
9e75ce23e7f165beca3195de487e8c1a880c9aeb53cad230e1173f606cd4f0d6 - Sigstore transparency entry: 1368351543
- Sigstore integration time:
-
Permalink:
muxi-ai/runtime@aba0425d9b63eb11a3dcc07fed2a58525e7afc01 -
Branch / Tag:
refs/heads/main - Owner: https://github.com/muxi-ai
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@aba0425d9b63eb11a3dcc07fed2a58525e7afc01 -
Trigger Event:
push
-
Statement type:
File details
Details for the file muxi_runtime-0.20260424.0-py3-none-any.whl.
File metadata
- Download URL: muxi_runtime-0.20260424.0-py3-none-any.whl
- Upload date:
- Size: 1.6 MB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.13
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
69b096a4e5d9955b73869e5a759ffd0e17fae53c6429fdb4e56a0da8fd707168
|
|
| MD5 |
b1254941927b150b48830d564a7c4d6e
|
|
| BLAKE2b-256 |
3db675dea60febd7d180a46cd26656440ea66fc1ef52b3bd132c04d512e749fa
|
Provenance
The following attestation bundles were made for muxi_runtime-0.20260424.0-py3-none-any.whl:
Publisher:
release.yml on muxi-ai/runtime
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
muxi_runtime-0.20260424.0-py3-none-any.whl -
Subject digest:
69b096a4e5d9955b73869e5a759ffd0e17fae53c6429fdb4e56a0da8fd707168 - Sigstore transparency entry: 1368351550
- Sigstore integration time:
-
Permalink:
muxi-ai/runtime@aba0425d9b63eb11a3dcc07fed2a58525e7afc01 -
Branch / Tag:
refs/heads/main - Owner: https://github.com/muxi-ai
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@aba0425d9b63eb11a3dcc07fed2a58525e7afc01 -
Trigger Event:
push
-
Statement type: