Run multiple LiveKit voice agents in a single shared worker process.
Project description
openrtc-python
Run N LiveKit voice agents in one worker. Pay the model-load cost once.
PyPI package name: openrtc.
Table of Contents
- The problem
- What openrtc does
- Installation
- Quick start: explicit registration with add()
- Quick start: one Python file per agent with discover()
- Memory: before and after
- Routing
- Greetings and session options
- Provider configuration
- CLI and TUI
- Public API at a glance
- Project structure
- Contributing
- License
The problem
You already ship three voice agents with livekit-agents. Each agent is its own worker on the same VPS. Every worker process loads the same shared stack: Python runtime, Silero VAD, and the turn-detection model. You are not loading three different models. You are loading the same stack three times because the process boundary forces it. On a 1–2 GB instance, that shows up as duplicate resident set for every idle worker. You pay RAM for copies you do not need.
What openrtc does
openrtc gives you one AgentPool in one worker: prewarm runs once, each incoming call still gets its own AgentSession, and you register multiple Agent subclasses on the pool so dispatch can pick one per session from metadata or fallbacks. This package does not replace your agent code. It does not sit between you and livekit.agents.Agent, @function_tool, RunContext, on_enter, on_exit, llm_node, stt_node, or tts_node. You keep your subclasses and tools as they are. You change how many workers you run, not how you write an agent.
Installation
OpenRTC requires Python 3.11 or newer. The LiveKit Silero / turn-detector
plugins depend on onnxruntime, which does not ship supported wheels for
Python 3.10 in current releases—use 3.11+ to avoid install failures.
pip install openrtc
The base install pulls in livekit-agents[openai,silero,turn-detector] so shared prewarm has the plugins it expects. The package ships a PEP 561 py.typed marker for downstream type checkers.
With uv (recommended in CONTRIBUTING.md):
uv add openrtc
uv add "openrtc[cli,tui]"
pip install 'openrtc[cli]'
Optional Textual sidecar for live metrics:
pip install 'openrtc[cli,tui]'
Set the same variables you use for any LiveKit worker:
export LIVEKIT_URL=ws://localhost:7880
export LIVEKIT_API_KEY=devkey
export LIVEKIT_API_SECRET=secret
For OpenAI-backed plugins, set OPENAI_API_KEY as you already do.
Quick start: explicit registration with add()
Use this when you want every agent registered in one place with explicit names and providers.
from livekit.agents import Agent
from livekit.plugins import openai
from openrtc import AgentPool
class RestaurantAgent(Agent):
def __init__(self) -> None:
super().__init__(instructions="You help callers make restaurant bookings.")
class DentalAgent(Agent):
def __init__(self) -> None:
super().__init__(instructions="You help callers manage dental appointments.")
pool = AgentPool()
pool.add(
"restaurant",
RestaurantAgent,
stt=openai.STT(model="gpt-4o-mini-transcribe"),
llm=openai.responses.LLM(model="gpt-4.1-mini"),
tts=openai.TTS(model="gpt-4o-mini-tts"),
greeting="Welcome to reservations.",
)
pool.add(
"dental",
DentalAgent,
stt=openai.STT(model="gpt-4o-mini-transcribe"),
llm=openai.responses.LLM(model="gpt-4.1-mini"),
tts=openai.TTS(model="gpt-4o-mini-tts"),
)
pool.run()
Quick start: one Python file per agent with discover()
Use this when you prefer one module per agent and optional @agent_config(...) on each class.
Create a directory (for example agents/) and add one .py file per agent. Then:
from pathlib import Path
from livekit.plugins import openai
from openrtc import AgentPool
pool = AgentPool(
default_stt=openai.STT(model="gpt-4o-mini-transcribe"),
default_llm=openai.responses.LLM(model="gpt-4.1-mini"),
default_tts=openai.TTS(model="gpt-4o-mini-tts"),
)
pool.discover(Path("./agents"))
pool.run()
Example file agents/restaurant.py:
from livekit.agents import Agent
from openrtc import agent_config
@agent_config(name="restaurant", greeting="Welcome to reservations.")
class RestaurantAgent(Agent):
def __init__(self) -> None:
super().__init__(instructions="You help callers make restaurant bookings.")
If a module has no @agent_config, the agent name defaults to the filename stem. STT, LLM, TTS, and greeting fall back to the pool defaults.
Discovered agents work with livekit dev and spawn-based workers on macOS. For add(), define agent classes at module scope so worker reload can import them.
Memory: before and after
Assume an illustrative ~400 MB idle baseline per worker for the shared stack (VAD, turn detector, and similar). Your measured RSS will differ by provider, model, and OS.
| Before openrtc | After openrtc | |
|---|---|---|
| Three workers, same stack | about 3 × 400 MB ≈ 1.2 GB idle baseline (three loads) | — |
| One worker, three registered agents | — | about one × 400 MB idle baseline (one load) plus per-session overhead |
Exact numbers depend on your providers, concurrency, and call patterns. The win is not loading that stack once per agent worker.
Routing
One process hosts several agent classes, so each session must resolve to a single registered name. AgentPool resolves the agent in this order:
ctx.job.metadata["agent"]ctx.job.metadata["demo"]ctx.room.metadata["agent"]ctx.room.metadata["demo"]- room name prefix match, such as
restaurant-call-123 - the first registered agent
If metadata names an agent that is not registered, you get a ValueError instead of a silent fallback.
Greetings and session options
You can pass a greeting and extra AgentSession options per registration.
pool.add(
"restaurant",
RestaurantAgent,
greeting="Welcome to reservations.",
session_kwargs={"turn_handling": {"interruption": {"enabled": False}}},
max_tool_steps=4,
preemptive_generation=True,
)
Direct keyword arguments win over the same keys inside session_kwargs.
By default, OpenRTC sets explicit turn_handling with the multilingual turn detector and VAD-based interruption. To opt into adaptive interruption, pass session_kwargs={"turn_handling": {"interruption": {"mode": "adaptive"}}}.
Provider configuration
Pass instantiated provider objects through to livekit-agents unchanged, for example:
openai.STT(model="gpt-4o-mini-transcribe")openai.responses.LLM(model="gpt-4.1-mini")openai.TTS(model="gpt-4o-mini-tts")
If you pass strings such as openai/gpt-4.1-mini, OpenRTC leaves them as-is and the LiveKit runtime interprets them for your deployment.
CLI and TUI
Install openrtc[cli] to get openrtc on your PATH. Subcommands follow the LiveKit Agents CLI shape (dev, start, console, connect, download-files), plus list and tui. For most commands you can pass the agents directory (or, for tui, the metrics JSONL file) as the first path argument instead of --agents-dir / --watch.
List what discovery would register (defaults are string passthroughs for livekit-agents, not constructed provider objects):
openrtc list \
./agents \
--default-stt openai/gpt-4o-mini-transcribe \
--default-llm openai/gpt-4.1-mini \
--default-tts openai/gpt-4o-mini-tts
Run a production worker (after exporting LIVEKIT_*):
openrtc start ./agents
Run a development worker:
openrtc dev ./agents
Same as openrtc dev --agents-dir ./agents. The metrics JSONL file is optional: add a second path only when you want JSONL output (same as --metrics-jsonl), e.g. openrtc dev ./agents ./openrtc-metrics.jsonl for openrtc tui.
Optional visibility: --dashboard prints a Rich summary in the terminal. --metrics-json-file ./runtime.json overwrites a JSON snapshot on each tick. Use that for scripts, dashboards, or CI. For JSON Lines plus a separate terminal UI, use --metrics-jsonl ./openrtc-metrics.jsonl on the worker and openrtc tui in another terminal (it tails ./openrtc-metrics.jsonl by default; override with --watch) after pip install 'openrtc[cli,tui]'.
Stable machine output: openrtc list --json and --plain. Combine --resources when you want footprint hints. OpenRTC-only flags are stripped before the handoff to LiveKit’s CLI parser.
Full flag lists live in docs/cli.md.
Public API at a glance
Everything openrtc exposes publicly is listed here. Anything else is internal and not treated as stable.
AgentPoolAgentConfigAgentDiscoveryConfigagent_config(...)ProviderValue— type alias for STT/LLM/TTS slot values (provider ID strings or LiveKit plugin instances)
On AgentPool:
add(...)discover(...)list_agents()get(name)remove(name)run()runtime_snapshot()drain_metrics_stream_events()— for JSONL export paths (mainly CLI; rare in app code)server
Project structure
src/openrtc/
├── __init__.py
├── py.typed
├── cli.py # lazy console entry / missing-extra hints
├── cli_app.py # Typer commands and programmatic main()
├── cli_types.py # shared CLI option aliases
├── cli_dashboard.py # Rich dashboard and list output
├── cli_reporter.py # background metrics reporter thread
├── cli_livekit.py # LiveKit argv/env handoff, pool run
├── cli_params.py # shared worker handoff option bundles
├── metrics_stream.py # JSONL metrics schema
├── provider_types.py # ProviderValue and related typing
├── tui_app.py # optional Textual sidecar
└── pool.py # AgentPool, discovery, routing
pool.py—AgentPool, discovery, routingcli.py/cli_app.py— Typer/Rich CLI (openrtc[cli])metrics_stream.py— JSONL metrics schematui_app.py— optional Textual sidecar (openrtc[tui])
Contributing
See CONTRIBUTING.md. CI runs Ruff and mypy on pull requests alongside the test suite.
License
MIT. See LICENSE.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file openrtc-0.0.17.tar.gz.
File metadata
- Download URL: openrtc-0.0.17.tar.gz
- Upload date:
- Size: 1.1 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
ff95c71ac510b38b7621dee8c074acee26b0d968b765c0e944436408522f52af
|
|
| MD5 |
17e3737ff50684a8ef5a7c3eea7e623c
|
|
| BLAKE2b-256 |
ff7b4ce2d9bd81ca3781f9dda3b76270f495e07fa49a130ddedb16690a1a09aa
|
File details
Details for the file openrtc-0.0.17-py3-none-any.whl.
File metadata
- Download URL: openrtc-0.0.17-py3-none-any.whl
- Upload date:
- Size: 41.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
dc6ba7d5638399c460983a4e180dd9129443a95f76e1e8af6474df6f535c3757
|
|
| MD5 |
78baeababc023110b0ddaa09f487c682
|
|
| BLAKE2b-256 |
3057d1a08d5349c57d1fe7ebb04a958f18c22cd68d78a5bd6441237989cf9885
|