Skip to main content

Pre-curated canonical memory + prose/code provenance for any AI coding client. MCP-native.

Project description

Librarian MCP

A real, measured alternative to "bigger context windows." Pre-curated canonical memory + prose/code provenance checking + benchmark metrics, delivered as a Model Context Protocol server that works across Claude Code, Cursor, VSCode (via Continue), and any MCP-capable client.

What it does

Five tools, all exposed via MCP:

Tool What it does Added
librarian_context Intent-aware canonical memory packet. Loads curated preload content scoped to your query intent (outreach, architecture, benchmark, founder voice, etc.). Eliminates the "forgets by prompt #21" failure mode. v0.1.0 (stub), v0.2.0 (intent-aware)
prose_provenance Deterministic drift detection between two document versions. Catches silently-removed voice anchors, stale canonical numbers, section changes, register shifts. v0.1.0
record_measurement Log a single benchmark measurement (vendor, model, condition, accuracy, cost, latency) to local JSONL. v0.2.0
metrics_summary Per-vendor and per-model aggregation of recorded measurements. Shows accuracy lift, cost savings, cache hit rate. v0.2.0
opt_in_share Toggle anonymous metrics sharing flag. Default OFF. Commons dashboard POST endpoint ships in a future release. v0.2.0

Why we built this

Independently measured result (Eyewitness Benchmark R10, April 2026, eight models across four vendors, 1,200 graded calls, inter-rater kappa 0.883/0.850):

  • Without the Librarian (COLD): mean 8.7% correct
  • With the Librarian (HOT): mean 94.8% correct — 86.1 percentage-point lift
  • Haiku 4.5 (cheapest) ties Opus 4.7 (most expensive) at 19x cost difference
  • 4.3x more right answers per dollar of compute

Applied inside Microsoft Copilot's inference path, the same architecture recovers an estimated $750M/year in waste. Inside Anthropic's developer tools, ~$130M/year. Full methodology in the R9 Empirical Test Companion Paper.

librarian_context — Intent API

librarian_context(intent="outreach", max_tokens=16000)
Intent What it loads Approx. tokens
"" (default) Base R9-v2 preload only ~4,500
"canonical" Base + canonical values + canonical laws ~15,000
"outreach" Base + canonical + Opening Gambit + letter queue + Cephas + Glass Door + Witness ~30,000
"architecture" Base + canonical + Pledge + IP split + Medallion + Pedestal Stake ~20,000
"founder_voice" Base + Rhetorical Keystones + Pine Books + Anachronism + Cloyd + Three-clock ~10,000
"benchmark" Base + R10 results + R9 brief + 75-Q bank + rubric + posture disclosure ~10,000
"operational" Union of outreach + canonical ~30,000

List inputs for union queries: intent='["benchmark", "founder_voice"]'

Returns:

{
  "packet": "...markdown...",
  "sections_included": ["r9v2_base.md", "canonical/canonical_values.yaml", ...],
  "token_count": 14832,
  "source_version": "a1b2c3d4e5f6",
  "truncation_note": null
}

metrics_summary — Schema

{
  "total_calls": 1200,
  "per_vendor": {
    "anthropic": {
      "calls": 600,
      "hot_accuracy": 95.3,
      "cold_baseline_est": 8.2,
      "dollars_saved_est": 42.17,
      "cache_hit_rate": 50.0
    }
  },
  "per_model": {
    "claude-haiku-4-5-20251001": { "..." : "..." }
  },
  "cumulative_hot_accuracy": 94.8,
  "cumulative_cold_baseline_est": 8.7,
  "cumulative_dollars_saved_est": 127.50,
  "opt_in_share": false,
  "since": "all_time"
}

Pricing

Tier Who it's for Price
Pledged Commons Any nonprofit, cooperative, academic institution, or public-service organization with IRS-verified EIN (or international equivalent) $0 forever. Full feature set. Under the Cooperative Defensive Patent Pledge.
Individual Single developer $0 (community edition, this repo) for local use; $15/mo for hosted multi-repo context + team sharing
Team 2–50 seats $10/seat/mo (min $50)
Enterprise 50+ seats, custom canonical schemas, audit logs, SAML, support Contact. Typically $50–100/seat/mo.

The commercial tiers pay for the commons. No grant funding, no VC, no extractive margin. Cost+20% on operating expense. That's it.

Why MCP (not a Cursor extension)

Because you shouldn't have to pick between your AI assistants. MCP servers work across Claude Code, Cursor (v0.45+), Continue (VSCode / JetBrains), Zed, and every MCP-capable client in the roadmap. One server, all your tools.

Install

Quick start (local, Python 3.10+)

git clone https://github.com/liana-banyan/librarian-mcp.git
cd librarian-mcp
pip install -e .
librarian-mcp  # starts on stdio for MCP clients

With optional dependencies

pip install -e ".[all]"   # tiktoken (accurate token counts) + anthropic + pyyaml
pip install -e ".[dev]"   # + pytest, ruff, mypy for development

Claude Code

claude mcp add librarian python -m librarian_mcp

Cursor

Add to ~/.cursor/mcp.json:

{
  "mcpServers": {
    "librarian": {
      "command": "python",
      "args": ["-m", "librarian_mcp"]
    }
  }
}

Continue (VSCode / JetBrains)

See docs/continue-integration.md.

Development

pip install -e ".[dev,all]"
ruff check src/ tests/          # lint
mypy --strict src/librarian_mcp/  # type check
pytest -v                        # test (34 tests)

Status

April 21, 2026 — v0.2.0. Intent-aware librarian_context live with bundled preload (R10-validated). Benchmark metrics recording live. Prose Provenance tool upgraded to v0.2.0. PyPI name librarian-mcp reserved. CI/CD staged.

License

AGPL-3.0. Commercial licensing for the paid tiers is a separate agreement; the Pledged Commons tier is covered by AGPL + the Cooperative Defensive Patent Pledge.

Contact


"You build the Features — We're building the Board."

Pledged into the commons. For the Keep.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

librarian_mcp-0.2.0.tar.gz (86.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

librarian_mcp-0.2.0-py3-none-any.whl (98.7 kB view details)

Uploaded Python 3

File details

Details for the file librarian_mcp-0.2.0.tar.gz.

File metadata

  • Download URL: librarian_mcp-0.2.0.tar.gz
  • Upload date:
  • Size: 86.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for librarian_mcp-0.2.0.tar.gz
Algorithm Hash digest
SHA256 b6c112c85dab072218542ea21ed31bf0938241ef2d1732c185a5a6f7012b8a8a
MD5 a062ccad7b2d1613a2d06090bddf9afa
BLAKE2b-256 61a18e191a647a66ac20f5029451f7338e1b647dc0d369a6f85402ff7b7cd006

See more details on using hashes here.

File details

Details for the file librarian_mcp-0.2.0-py3-none-any.whl.

File metadata

  • Download URL: librarian_mcp-0.2.0-py3-none-any.whl
  • Upload date:
  • Size: 98.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for librarian_mcp-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 345ec7d7771cbc47d670d7f52b3dd24c04f86e0e8308a8debc20f6211a800d98
MD5 7d3289ad684a62f347f25f621a1af0ac
BLAKE2b-256 1c8d20293a3d43b3483d7e1ae25ba50dc26bde13586b3ec5f21f528579a58e7b

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page