Local LLM router that cuts premium-model spend with adaptive 3-tier routing, OpenAI + Anthropic compatible
Project description
English | 简体中文
UncommonRoute
Cut your LLM costs by 82% with automatic model routing.
Most of your LLM budget goes to simple tasks that don't need a premium model. UncommonRoute picks the cheapest model that still gets the job done — automatically.
Quick Start
1. Install
pipx install uncommon-route
pipx is the best default for most CLI users: it installs UncommonRoute into its own isolated environment, keeps your system Python clean, and gives you a clean uninstall path.
If you do not have pipx yet, prefer your OS package manager when it is available (brew install pipx on macOS, sudo apt install pipx on recent Ubuntu, sudo dnf install pipx on Fedora), then run pipx ensurepath.
If that is not available, see the pipx installation guide or install it with:
python3 -m pip install --user pipx
python3 -m pipx ensurepath
If you already work inside a virtual environment, pip is still fine:
python3 -m pip install uncommon-route
Install troubleshooting: pip vs. pipx
- If you are installing a command-line app for everyday use, prefer
pipx install uncommon-route. - If you are already inside a project virtual environment, use
python -m pip install uncommon-routeinside that environment. - Prefer
python3 -m pip ...over barepip ...when you are unsure which Python interpreterpippoints at. - If your OS Python reports an "externally managed environment" error, use
pipxor a virtual environment instead of forcing a system-wide install. - If you need a specific interpreter,
pipxcan target it directly, for example:pipx install --python python3.12 uncommon-route.
2. Run the guided setup
uncommon-route init
The wizard walks you through:
- choosing a connection path: Commonstack, local/custom upstream, or BYOK
- saving upstream credentials locally
- configuring Claude Code, Codex, or OpenAI SDK / Cursor
- optionally starting the proxy in background
If you prefer to sanity-check before starting the proxy:
uncommon-route doctor
3. Point your client at the proxy
| Client | Change |
|---|---|
| Claude Code | export ANTHROPIC_BASE_URL="http://localhost:8403" and export ANTHROPIC_AUTH_TOKEN="not-needed" |
| Codex / Cursor / OpenAI SDK | export OPENAI_BASE_URL="http://localhost:8403/v1" |
| OpenClaw | Plugin — see openclaw.ai |
Then use uncommon-route/auto as the model ID:
client = OpenAI(base_url="http://localhost:8403/v1")
resp = client.chat.completions.create(model="uncommon-route/auto", messages=msgs)
# → simple tasks → cheap model, complex tasks → premium model
Works with Claude Code, Codex, Cursor, the OpenAI SDK, and OpenClaw.
Manual setup (advanced)
Commonstack managed upstream
export UNCOMMON_ROUTE_UPSTREAM="https://api.commonstack.ai/v1"
export UNCOMMON_ROUTE_API_KEY="csk-your-key"
uncommon-route serve
One key gives you OpenAI, Anthropic, Google, xAI, MiniMax, Moonshot, DeepSeek, and more — consolidated billing, no per-provider setup.
Bring your own keys (BYOK)
uncommon-route provider add openai sk-...
uncommon-route provider add anthropic sk-ant-...
uncommon-route provider add google AIza...
# also supported: xai, minimax, moonshot, deepseek
uncommon-route serve
Auto-routing will only consider models backed by a registered provider.
Note: UncommonRoute does not auto-read
OPENAI_API_KEY/ANTHROPIC_API_KEY. Useuncommon-route init, a saved connection, or one of the manual paths above.
How It Works
Every request is analyzed by three independent signals, then routed to the cheapest capable model:
"hello" → 🟢 nano $0.0008
"fix the typo on line 3" → 🟢 deepseek $0.0012
"refactor this 500-line module" → 🟠 sonnet $0.0337
"design a distributed scheduler" → 🔴 opus $0.0562
| Signal | What it does | Speed (CPU, warm) |
|---|---|---|
| Metadata | Conversation structure, tool usage, depth | <1ms |
| Embedding | Semantic similarity to known task patterns (bge-small) | ~20ms |
| Structural | Text complexity features (shadow mode) | <1ms |
End-to-end route() overhead on a warm process is ~20–25ms (dominated by the embedding signal). Cold start is a few hundred ms for the first request. GPU or a cached embedding path can bring this under 5ms.
Signals vote. The ensemble picks the tier. The router selects the cheapest model in that tier. If uncertain, it leans conservative — better to spend a little more than to fail the task.
It gets smarter over time. Signal weights adjust from routing outcomes. The embedding index grows with usage. Low-confidence predictions automatically escalate.
Why v2
Our v1 classifier hit 88.5% accuracy on clean benchmark data. We shipped it.
Then we tested on real agent conversations — multi-turn, tool-calling, messy context — and accuracy dropped to 43%. More than half the routing decisions were wrong.
We didn't patch it. We rebuilt from scratch.
| v1 | v2 | |
|---|---|---|
| Accuracy | 43% | 78% |
| Task pass rate | 100% (cheated — always chose most expensive) | 93.4% (real routing) |
| Cost savings | 0% | 82% |
We're telling you this because we'd rather you trust our numbers than be impressed by them.
Benchmarks
Tested on CommonRouterBench — 970 real agent task traces across SWE-Bench, BFCL, MT-RAG, QMSum, and PinchBench. All numbers measured end-to-end through the production code path.
| Metric | Value |
|---|---|
| Cost savings | 82% vs always-premium |
| Task pass rate | 93.4% |
| Routing overhead | ~20–25ms (warm process, CPU, bge-small embedding) |
| Accuracy | 78% tier match |
python scripts/eval_v2.py # reproduce it yourself
Dashboard
uncommon-route serve
# → http://localhost:8403/dashboard/
Real-time monitoring, interactive playground, cost tracking, and model routing configuration — all in a Nothing Design-inspired interface.
Diagnostics
When a user hits a routing or upstream issue, you can export a local support bundle without guessing which logs to collect:
uncommon-route support bundle
uncommon-route support request <request_id>
The bundle includes recent request traces, recent errors, stats summaries, provider/config snapshots, and redacted local state. It stays on your machine until you choose to share it.
Stopping and Uninstalling
To stop the proxy:
- foreground run: press
Ctrl+Cin the terminal runninguncommon-route serve - background daemon: run
uncommon-route stop - background logs: run
uncommon-route logs --follow
To stop routing your clients through UncommonRoute, remove or comment out the shell block that uncommon-route init added to your shell rc file (~/.zshrc, ~/.bashrc, or ~/.config/fish/config.fish), then restart your terminal. For the current shell only, you can also unset the proxy variables:
unset OPENAI_BASE_URL OPENAI_API_KEY ANTHROPIC_BASE_URL ANTHROPIC_AUTH_TOKEN ANTHROPIC_API_KEY
To uninstall the package:
pipx uninstall uncommon-route
# or, if you installed it with pip in a specific environment:
python3 -m pip uninstall uncommon-route
If you also want to remove local state, delete ~/.uncommon-route/. That directory contains saved connections, provider keys, logs, traces, and support bundles.
Configuration
Routing modes
| Mode | Model ID | Behavior |
|---|---|---|
| auto | uncommon-route/auto |
Balanced — best quality-per-dollar |
| fast | uncommon-route/fast |
Cost-first — cheapest acceptable |
| best | uncommon-route/best |
Quality-first — strongest available |
Spend limits
uncommon-route spend set daily 20.00
uncommon-route spend status
Managing providers
uncommon-route provider list
uncommon-route provider add <name> <api-key>
uncommon-route provider remove <name>
Supported names: commonstack, openai, anthropic, google, xai, minimax, moonshot, deepseek. See Quick Start for the two setup paths (managed upstream vs. BYOK).
All environment variables
| Variable | Meaning |
|---|---|
UNCOMMON_ROUTE_UPSTREAM |
Upstream base URL for the managed path (e.g. https://api.commonstack.ai/v1). Ignored in BYOK mode. |
UNCOMMON_ROUTE_API_KEY |
API key paired with UNCOMMON_ROUTE_UPSTREAM. Not a fallback for per-provider keys. |
UNCOMMON_ROUTE_PORT |
Local proxy port (default 8403) |
Privacy
Runs entirely on your machine. No data leaves unless you opt in.
uncommon-route telemetry status
Diagnostics exports are also local-first: uncommon-route support bundle writes a redacted zip under ~/.uncommon-route/support/ by default.
Development
git clone https://github.com/CommonstackAI/UncommonRoute.git
cd UncommonRoute && pip install -e ".[dev]"
python -m pytest tests -v
License
MIT — see LICENSE.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file uncommon_route-0.7.14.tar.gz.
File metadata
- Download URL: uncommon_route-0.7.14.tar.gz
- Upload date:
- Size: 2.0 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
933f2e67785c0c0a7cedbf9b322fa48ae6b20c8c7cd05ec07e8df66dc46a510a
|
|
| MD5 |
150c4e7a86e3a99712c43765de88f841
|
|
| BLAKE2b-256 |
fc541fdd3db1af05d93e31b0e8bb8f52afef864c33931fc681c411dd0f0f4387
|
File details
Details for the file uncommon_route-0.7.14-py3-none-any.whl.
File metadata
- Download URL: uncommon_route-0.7.14-py3-none-any.whl
- Upload date:
- Size: 2.0 MB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
c72baa78091e9cba498aea660d240f8270552bb27b151d2efe9b00ee1d72b189
|
|
| MD5 |
62d91faa562a3b3554495c237cbcd4b6
|
|
| BLAKE2b-256 |
116ea9ac1999f172f5b889b8b81c8688469882fcba82286104d43d19eb775f67
|