Skip to main content

Unified MCP server for AI-assisted development — on-demand Obsidian vault access + worker delegation

Project description

hive-vault

CI PyPI Python 3.12+ Docs License: MIT

Your AI coding assistant forgets everything between sessions. Hive fixes that.

Every session, your assistant loads 800+ lines of static context. Most of it is irrelevant. You pay the full token cost every time. And next session? It starts from zero again.

Hive is an MCP server that connects your AI assistant to an Obsidian vault. Instead of loading everything upfront, it queries only what's needed — architecture decisions, lessons learned, project context — all on demand via MCP.

The numbers:

Metric Without Hive With Hive
Context loaded per session ~800 lines (static) ~50 lines (on demand)
Token cost for context 100% every session 6% average per query
Knowledge retained between sessions 0% 100% (in vault)
Time to find past decisions Manual search vault_search in seconds

Measured on a real vault with 19 projects, 200+ files. See benchmarks.

Install (30 seconds)

One command. No cloning, no venv, no config files. Use user scope (-s user) so Hive works across all your projects — that's where cross-project knowledge shines.

Claude Code:

claude mcp add -s user hive -- uvx --upgrade hive-vault

Gemini CLI:

gemini mcp add -s user hive-vault uvx -- --upgrade hive-vault

OpenAI Codex CLI — add to ~/.codex/config.toml:

[mcp_servers.hive-vault]
command = "uvx"
args = ["--upgrade", "hive-vault"]

GitHub Copilot (VS Code) — add to .vscode/mcp.json:

{
  "servers": {
    "hive-vault": {
      "command": "uvx",
      "args": ["--upgrade", "hive-vault"]
    }
  }
}

Other MCP clients (Cursor, Windsurf, etc.): point your client at uvx --upgrade hive-vault via stdio transport.

Then ask your assistant:

"Use vault_list_projects to see my vault"

That's it. You're running.

What You Get

16 Vault Tools — your knowledge, on demand

Tool What it does
vault_query Load project context, tasks, roadmap, lessons — or any file by path
vault_search Full-text search with metadata filters and regex support
vault_smart_search Ranked results with relevance scoring (status + recency + match density)
session_briefing One call = tasks + lessons + git log + health. Start every session here
vault_list_projects See all projects in your vault
vault_list_files Browse project structure with glob pattern filtering
vault_health File counts, staleness metrics, coverage gaps per project
vault_validate Drift detector — find broken frontmatter, stale files, broken wikilinks
vault_recent What changed in the last N days (via git + frontmatter)
vault_update Write to vault with YAML validation + auto git commit
vault_create Create files with auto-generated frontmatter + auto git commit
vault_patch Surgical find-and-replace with ambiguity rejection + auto git commit
capture_lesson Capture a lesson inline — deduplicates, appends to 90-lessons.md
extract_lessons Worker-powered batch extraction — send text, get structured lessons
vault_summarize Small files returned directly, large files delegated for compression
vault_usage Tool call analytics — which tools, which projects, how many tokens

3 Worker Tools — delegate to cheaper models

Tool What it does
delegate_task Route tasks to Ollama (free, local) or OpenRouter (free/paid cloud)
list_models See all available models across providers
worker_status Budget remaining, connectivity, usage stats

Routing: Ollama first (free) → OpenRouter free tier → OpenRouter paid ($1/mo cap) → reject.

Your primary model handles architecture. Cheaper models handle boilerplate. extract_lessons uses workers to batch-extract lessons from session notes — saving your primary model's tokens.

Before / After

Before Hive — static CLAUDE.md:

# My Project
## Architecture
[200 lines of decisions you made 3 months ago]
## Standards
[150 lines of coding patterns]
## Lessons
[100 lines of past bugs]
## Tasks
[50 lines of backlog]
# ...loaded every single session, whether relevant or not

With Hive — dynamic, on demand:

# Only when the assistant needs architecture context:
vault_query(project="my-project", section="context")

# Only when searching for a past decision:
vault_search(query="database migration strategy")

# Start of session — just the essentials:
session_briefing(project="my-project")

Configure Your Vault

Default vault path: ~/Projects/knowledge. To change it:

# Claude Code
claude mcp add -s user hive -e VAULT_PATH=/path/to/vault -- uvx --upgrade hive-vault

# Gemini CLI
gemini mcp add -s user -e VAULT_PATH=/path/to/vault hive-vault uvx -- --upgrade hive-vault

Enable Worker Delegation (optional)

claude mcp add -s user hive \
  -e VAULT_PATH=/path/to/vault \
  -e HIVE_OLLAMA_ENDPOINT=http://your-ollama:11434 \
  -e OPENROUTER_API_KEY=sk-or-... \
  -- uvx --upgrade hive-vault

All Configuration

Variable Default Description
VAULT_PATH ~/Projects/knowledge Path to your Obsidian vault
HIVE_OLLAMA_ENDPOINT http://localhost:11434 Ollama API endpoint
HIVE_OLLAMA_MODEL qwen2.5-coder:7b Default Ollama model
HIVE_OPENROUTER_API_KEY OpenRouter API key (also reads OPENROUTER_API_KEY)
HIVE_OPENROUTER_MODEL qwen/qwen3-coder:free Default free tier model
HIVE_OPENROUTER_PAID_MODEL qwen/qwen3-coder Paid tier model
HIVE_OPENROUTER_BUDGET 1.0 Monthly budget cap (USD)
HIVE_VAULT_SCOPES {"projects": "10_projects", "meta": "00_meta"} JSON mapping of scope names to vault subdirectories

See full configuration reference for all 15 environment variables.

Recommended Workflow

The highest-value setup combines three tools:

  1. Obsidian — local-first knowledge base with 1M+ community, Markdown native, no lock-in
  2. Obsidian Git — auto-commits your vault changes on a schedule (version history for free)
  3. Hive — bridges your vault to any AI coding assistant via MCP

Your assistant writes lessons and decisions to the vault → Obsidian Git auto-commits → next session, everything is there. No manual sync. No context lost.

Hive works with any directory of Markdown files — Obsidian is recommended, not required.

Vault Structure

For best results, follow this layout:

~/Projects/knowledge/
├── 00_meta/patterns/          # cross-project patterns
├── 10_projects/
│   ├── my-project/
│   │   ├── 00-context.md      # vault_query section="context"
│   │   ├── 10-roadmap.md      # vault_query section="roadmap"
│   │   ├── 11-tasks.md        # vault_query section="tasks"
│   │   ├── 90-lessons.md      # vault_query section="lessons"
│   │   └── 30-architecture/   # any path works with vault_query path="..."
│   └── another-project/
└── ...

Make Your Assistant Use Hive Consistently

MCP tools don't activate on their own. Add this to your project's CLAUDE.md (or equivalent):

## Vault & Knowledge (Hive MCP)

When hive-vault MCP is available:
- `session_briefing(project="myproject")` — start every session here
- `vault_query(project="myproject", section="context")` — project overview
- `vault_search(query="...")` — find past decisions
- `capture_lesson(...)` — capture insights inline, don't wait until session end

Without these instructions, your assistant uses Hive inconsistently. With them, it uses Hive every session, predictably.

Resources & Prompts

5 MCP Resources for auto-discoverable data:

URI Description
hive://projects All vault projects with file counts
hive://health Vault health metrics
hive://projects/{project}/context Project context
hive://projects/{project}/tasks Project backlog
hive://projects/{project}/lessons Lessons learned

4 MCP Prompts for guided workflows:

Prompt Description
retrospective End-of-session review → extract lessons to vault
delegate Structured protocol for worker delegation
vault_sync Post-sprint vault sync — reconcile docs with shipped code
benchmark Estimate token savings from Hive in the current session

Architecture

MCP Host (Claude Code, Gemini CLI, Codex CLI, Cursor, ...)
    └── hive-vault (MCP server, stdio)
            ├── Vault Tools (16) ── Obsidian vault (Markdown + YAML frontmatter)
            │     query, search, smart_search, list_files, patch,
            │     update, create, capture_lesson, extract_lessons,
            │     summarize, validate, session_briefing, recent,
            │     usage, health, list_projects
            │
            └── Worker Tools (3) ── Task delegation + routing:
                  delegate_task        1. Ollama (local, free)
                  list_models          2. OpenRouter free tier
                  worker_status        3. OpenRouter paid ($1/mo cap)
                                       4. Reject → host handles it

Development

See CONTRIBUTING.md for setup, code standards, and PR workflow.

git clone https://github.com/mlorentedev/hive.git
cd hive
make install   # create venv + install deps
make check     # lint + typecheck + test (265 tests, 92% coverage)

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

hive_vault-1.7.2.tar.gz (216.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

hive_vault-1.7.2-py3-none-any.whl (33.5 kB view details)

Uploaded Python 3

File details

Details for the file hive_vault-1.7.2.tar.gz.

File metadata

  • Download URL: hive_vault-1.7.2.tar.gz
  • Upload date:
  • Size: 216.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for hive_vault-1.7.2.tar.gz
Algorithm Hash digest
SHA256 e6ec4fdc62afc2e89c8a0fd44081d07877edb45d66c378e09707335bcf9fd237
MD5 cbb0b690397510305943277c78c4813d
BLAKE2b-256 dd7116bc3ec6298b947e52edc99582ca83ada9cee8fe490522880378f24394a7

See more details on using hashes here.

Provenance

The following attestation bundles were made for hive_vault-1.7.2.tar.gz:

Publisher: release.yml on mlorentedev/hive

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file hive_vault-1.7.2-py3-none-any.whl.

File metadata

  • Download URL: hive_vault-1.7.2-py3-none-any.whl
  • Upload date:
  • Size: 33.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for hive_vault-1.7.2-py3-none-any.whl
Algorithm Hash digest
SHA256 426c98388411865e0b76fb5156daf1d835632e1131ff04dfd9bbfc6346582d95
MD5 a2cb87106006c5f316e2853f7dde3bd0
BLAKE2b-256 80e8a6bef3e6e197eaee0f5cd176618f8ab1ef58392d69819ac23fd486477fe2

See more details on using hashes here.

Provenance

The following attestation bundles were made for hive_vault-1.7.2-py3-none-any.whl:

Publisher: release.yml on mlorentedev/hive

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page