Pre-curated canonical memory + prose/code provenance for any AI coding client. MCP-native.
Project description
Librarian MCP
A real, measured alternative to "bigger context windows." Pre-curated canonical memory + prose/code provenance checking + benchmark metrics, delivered as a Model Context Protocol server that works across Claude Code, Cursor, VSCode (via Continue), and any MCP-capable client.
What it does
Five tools, all exposed via MCP:
| Tool | What it does | Added |
|---|---|---|
librarian_context |
Intent-aware canonical memory packet. Loads curated preload content scoped to your query intent (outreach, architecture, benchmark, founder voice, etc.). Eliminates the "forgets by prompt #21" failure mode. | v0.1.0 (stub), v0.2.0 (intent-aware) |
prose_provenance |
Deterministic drift detection between two document versions. Catches silently-removed voice anchors, stale canonical numbers, section changes, register shifts. | v0.1.0 |
record_measurement |
Log a single benchmark measurement (vendor, model, condition, accuracy, cost, latency) to local JSONL. | v0.2.0 |
metrics_summary |
Per-vendor and per-model aggregation of recorded measurements. Shows accuracy lift, cost savings, cache hit rate. | v0.2.0 |
opt_in_share |
Toggle anonymous metrics sharing flag. Default OFF. Commons dashboard POST endpoint ships in a future release. | v0.2.0 |
Why we built this
Independently measured result (Eyewitness Benchmark R10, April 2026, eight models across four vendors, 1,200 graded calls, inter-rater kappa 0.883/0.850):
- Without the Librarian (COLD): mean 8.7% correct
- With the Librarian (HOT): mean 94.8% correct — 86.1 percentage-point lift
- Haiku 4.5 (cheapest) ties Opus 4.7 (most expensive) at 19x cost difference
- 4.3x more right answers per dollar of compute
Applied inside Microsoft Copilot's inference path, the same architecture recovers an estimated $750M/year in waste. Inside Anthropic's developer tools, ~$130M/year. Full methodology in the R9 Empirical Test Companion Paper.
librarian_context — Intent API
librarian_context(intent="outreach", max_tokens=16000)
| Intent | What it loads | Approx. tokens |
|---|---|---|
"" (default) |
Base R9-v2 preload only | ~4,500 |
"canonical" |
Base + canonical values + canonical laws | ~15,000 |
"outreach" |
Base + canonical + Opening Gambit + letter queue + Cephas + Glass Door + Witness | ~30,000 |
"architecture" |
Base + canonical + Pledge + IP split + Medallion + Pedestal Stake | ~20,000 |
"founder_voice" |
Base + Rhetorical Keystones + Pine Books + Anachronism + Cloyd + Three-clock | ~10,000 |
"benchmark" |
Base + R10 results + R9 brief + 75-Q bank + rubric + posture disclosure | ~10,000 |
"operational" |
Union of outreach + canonical |
~30,000 |
List inputs for union queries: intent='["benchmark", "founder_voice"]'
Returns:
{
"packet": "...markdown...",
"sections_included": ["r9v2_base.md", "canonical/canonical_values.yaml", ...],
"token_count": 14832,
"source_version": "a1b2c3d4e5f6",
"truncation_note": null
}
metrics_summary — Schema
{
"total_calls": 1200,
"per_vendor": {
"anthropic": {
"calls": 600,
"hot_accuracy": 95.3,
"cold_baseline_est": 8.2,
"dollars_saved_est": 42.17,
"cache_hit_rate": 50.0
}
},
"per_model": {
"claude-haiku-4-5-20251001": { "..." : "..." }
},
"cumulative_hot_accuracy": 94.8,
"cumulative_cold_baseline_est": 8.7,
"cumulative_dollars_saved_est": 127.50,
"opt_in_share": false,
"since": "all_time"
}
Pricing
| Tier | Who it's for | Price |
|---|---|---|
| Pledged Commons | Any nonprofit, cooperative, academic institution, or public-service organization with IRS-verified EIN (or international equivalent) | $0 forever. Full feature set. Under the Cooperative Defensive Patent Pledge. |
| Individual | Single developer | $0 (community edition, this repo) for local use; $15/mo for hosted multi-repo context + team sharing |
| Team | 2–50 seats | $10/seat/mo (min $50) |
| Enterprise | 50+ seats, custom canonical schemas, audit logs, SAML, support | Contact. Typically $50–100/seat/mo. |
The commercial tiers pay for the commons. No grant funding, no VC, no extractive margin. Cost+20% on operating expense. That's it.
Why MCP (not a Cursor extension)
Because you shouldn't have to pick between your AI assistants. MCP servers work across Claude Code, Cursor (v0.45+), Continue (VSCode / JetBrains), Zed, and every MCP-capable client in the roadmap. One server, all your tools.
Install
Quick start (local, Python 3.10+)
git clone https://github.com/liana-banyan/librarian-mcp.git
cd librarian-mcp
pip install -e .
librarian-mcp # starts on stdio for MCP clients
With optional dependencies
pip install -e ".[all]" # tiktoken (accurate token counts) + anthropic + pyyaml
pip install -e ".[dev]" # + pytest, ruff, mypy for development
Claude Code
claude mcp add librarian python -m librarian_mcp
Cursor
Add to ~/.cursor/mcp.json:
{
"mcpServers": {
"librarian": {
"command": "python",
"args": ["-m", "librarian_mcp"]
}
}
}
Continue (VSCode / JetBrains)
See docs/continue-integration.md.
Development
pip install -e ".[dev,all]"
ruff check src/ tests/ # lint
mypy --strict src/librarian_mcp/ # type check
pytest -v # test (34 tests)
Status
April 21, 2026 — v0.2.0. Intent-aware librarian_context live with bundled preload (R10-validated). Benchmark metrics recording live. Prose Provenance tool upgraded to v0.2.0. PyPI name librarian-mcp reserved. CI/CD staged.
License
AGPL-3.0. Commercial licensing for the paid tiers is a separate agreement; the Pledged Commons tier is covered by AGPL + the Cooperative Defensive Patent Pledge.
Contact
- General: hello@liana-banyan.com
- Enterprise: enterprise@liana-banyan.com
- Press / AI policy / datacenter-alternative questions: press@liana-banyan.com
- Founder: Jonathan Jones, Founder & General Manager, Liana Banyan Corporation (Wyoming C-Corp)
"You build the Features — We're building the Board."
Pledged into the commons. For the Keep.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file librarian_mcp-0.2.0.tar.gz.
File metadata
- Download URL: librarian_mcp-0.2.0.tar.gz
- Upload date:
- Size: 86.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
b6c112c85dab072218542ea21ed31bf0938241ef2d1732c185a5a6f7012b8a8a
|
|
| MD5 |
a062ccad7b2d1613a2d06090bddf9afa
|
|
| BLAKE2b-256 |
61a18e191a647a66ac20f5029451f7338e1b647dc0d369a6f85402ff7b7cd006
|
File details
Details for the file librarian_mcp-0.2.0-py3-none-any.whl.
File metadata
- Download URL: librarian_mcp-0.2.0-py3-none-any.whl
- Upload date:
- Size: 98.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
345ec7d7771cbc47d670d7f52b3dd24c04f86e0e8308a8debc20f6211a800d98
|
|
| MD5 |
7d3289ad684a62f347f25f621a1af0ac
|
|
| BLAKE2b-256 |
1c8d20293a3d43b3483d7e1ae25ba50dc26bde13586b3ec5f21f528579a58e7b
|