Cost-aware model routing for LangChain agents based on task phase
Project description
langchain-router
Your agent doesn't need the expensive model for every call.
Most calls are just the model picking which file to read next or which pattern to search for. A smaller model does that fine. This middleware detects when the agent is doing that kind of work and routes to a fast model automatically.
Quick Install
pip install langchain-router
🤔 What is this?
Agent sessions have a pattern. The user says something, the agent thinks about it (planning). Then it reads files, searches code, runs commands (execution). Sometimes something breaks (recovery). Then the user says something again.
Planning and recovery need the primary model. Execution doesn't. RouterMiddleware detects which phase the agent is in and routes accordingly.
| What just happened | Phase | Model |
|---|---|---|
| User spoke | planning | primary |
| Tool call succeeded | execution | fast |
| Tool call failed | recovery | primary |
One status check, two isinstance checks. No ML. No embeddings. No classifier.
from langchain.agents import create_agent
from langchain_router import RouterMiddleware
agent = create_agent(
model="anthropic:claude-sonnet-4-6",
tools=[...],
middleware=[RouterMiddleware(fast="anthropic:claude-haiku-4-5-20251001")],
)
With CollapseMiddleware
from langchain_collapse import CollapseMiddleware
middleware = [
CollapseMiddleware(),
RouterMiddleware(fast="anthropic:claude-haiku-4-5-20251001"),
]
flowchart TB
A["📥 37 messages"] --> B["CollapseMiddleware"]
B --> C["📥 9 messages"]
C --> D["RouterMiddleware"]
D --> E{"phase?"}
E --> |"execution · 83%"| F["⚡ Haiku"]
E --> |"planning"| G["🧠 Sonnet"]
E --> |"recovery"| G
style A fill:#ff6b6b,stroke:#e03131,color:#fff
style B fill:#339af0,stroke:#1c7ed6,color:#fff
style C fill:#339af0,stroke:#1c7ed6,color:#fff
style D fill:#51cf66,stroke:#2f9e44,color:#fff
style E fill:#fff3bf,stroke:#f59f00,color:#333
style F fill:#20c997,stroke:#099268,color:#fff
style G fill:#845ef7,stroke:#7048e8,color:#fff
Verified with real API calls
81 real API calls across 9 model pairs, 4 providers. Each pair ran a 9-call agent session: planning, 6 execution turns, recovery from a failed tool call, and re-planning. Every call routed correctly. Full per-call details.
| Primary | Fast | Provider | Calls | Routing |
|---|---|---|---|---|
| Claude Opus 4.6 | Claude Haiku 4.5 | Anthropic → Anthropic | 9 | 9/9 ✓ |
| Claude Opus 4.6 | GLM-5 | Anthropic → Z.ai | 9 | 9/9 ✓ |
| Claude Opus 4.6 | MiniMax M2.7 | Anthropic → MiniMax | 9 | 9/9 ✓ |
| Claude Sonnet 4.6 | GLM-5 | Anthropic → Z.ai | 9 | 9/9 ✓ |
| Claude Sonnet 4.6 | MiniMax M2.7 | Anthropic → MiniMax | 9 | 9/9 ✓ |
| GPT-5.4 | GPT-4.1-mini | OpenAI → OpenAI | 9 | 9/9 ✓ |
| GPT-5.4 | GLM-5 | OpenAI → Z.ai | 9 | 9/9 ✓ |
| GPT-5.4 | MiniMax M2.7 | OpenAI → MiniMax | 9 | 9/9 ✓ |
| GLM-5 | MiniMax M2.7 | Z.ai → MiniMax | 9 | 9/9 ✓ |
Patterns tested: same-provider (Anthropic→Anthropic, OpenAI→OpenAI), cross-provider (Anthropic→Z.ai, OpenAI→MiniMax), frontier→open (Opus→GLM-5, GPT-5.4→MiniMax), open→open (GLM-5→MiniMax).
Cost impact
In a typical coding session (8 file reads, 4 greps, 1 edit, 1 test failure, 1 fix, 1 re-test, 2 user messages), 83% of model calls are execution phase and route to the fast model.
| Primary | Fast | Per session | Saved | Annual (10 devs, 20/day) |
|---|---|---|---|---|
| Opus 4.6 | MiniMax M2.7 | $4.19 → $0.75 | 82% | $171,938/yr |
| Opus 4.6 | GLM-5 | $4.19 → $0.80 | 81% | $169,125/yr |
| Opus 4.6 | Haiku 4.5 | $4.19 → $0.88 | 79% | $165,075/yr |
| Sonnet 4.6 | MiniMax M2.7 | $0.84 → $0.19 | 78% | $32,437/yr |
| GPT-5.4 | MiniMax M2.7 | $0.77 → $0.18 | 77% | $29,438/yr |
| GPT-5.4 | GPT-4.1-mini | $0.77 → $0.21 | 72% | $27,675/yr |
| Sonnet 4.6 | GLM-5 | $0.84 → $0.24 | 71% | $29,625/yr |
| GPT-5.4 | GLM-5 | $0.77 → $0.23 | 70% | $26,625/yr |
| GLM-5 | MiniMax M2.7 | $0.13 → $0.07 | 45% | $2,813/yr |
Token counts assumed for cost projection (8k input, 1.5k output per call). Pricing as of April 2026. See benchmark for methodology.
On false positives
Recovery detection checks ToolMessage.status first (set by the agent when a tool call fails), then falls back to content scanning for error, traceback, exception, failed. Code containing those words (like def handle_error) routes to the primary model. That's the safe direction: more capability than needed, never less.
📖 Documentation
- Source (single file, ~170 lines)
- Benchmark (cost projections across providers)
- Verified Results (81 real API calls, 9 model pairs)
- Tests (unit, integration, and property based invariant tests)
💁 Contributing
git clone https://github.com/johanity/langchain-router.git
cd langchain-router
pip install -e ".[test]"
pytest
📕 License
MIT
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file langchain_router-0.1.1.tar.gz.
File metadata
- Download URL: langchain_router-0.1.1.tar.gz
- Upload date:
- Size: 16.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
035566d11e9b67f19547b8b51d10f11fbbf31f11e63f57dfba2ef108237cddf6
|
|
| MD5 |
5df274e3326071d4e6069870a7d45c51
|
|
| BLAKE2b-256 |
6f9e10e3bf4fcb66afcf2bd2fcd9ac29182bb771b921a2ed2e9f9d2820785b84
|
File details
Details for the file langchain_router-0.1.1-py3-none-any.whl.
File metadata
- Download URL: langchain_router-0.1.1-py3-none-any.whl
- Upload date:
- Size: 6.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
940746268343ac31746f2ed72f66d172c1d9bddc4e730b0a34d0bae234ad8d7e
|
|
| MD5 |
393d8f8bed24cc98b270ee21e3f84d2f
|
|
| BLAKE2b-256 |
230f98a8a756fc6bc2ffe583f1c4fb2d5b3417498326974326bcddc98a493d83
|