Skip to main content

MCP server that turns an Obsidian vault into semantic memory for any coding agent

Project description

second-brain-mcp

PyPI version Python versions License: MIT CI

Turn your Obsidian vault into semantic memory for any MCP-capable coding agent.

v1.0.0 — early release. Works end-to-end on macOS and Linux; published on PyPI on 2026-04-15. Windows is untested. Feedback, bug reports, and testing notes are very welcome — please open an issue if anything misbehaves or if the docs are unclear.

What it does. Ships a stdio MCP server with four read-only tools (obsidian_overview, obsidian_search, obsidian_read, obsidian_backlinks) plus one prompt (to_obsidian) that drives a human-in-the-loop write workflow. Works with Claude Code, Cursor, Zed, or any client that speaks MCP.

What it is not. No auto-writes — nothing is written to the vault without an explicit /to_obsidian invocation. No background extraction, no reflection or compaction loops running on your behalf. The LLM does the extraction work when you call /to_obsidian, but every candidate note requires your explicit per-note approval before it is written. You control what enters the vault; the server only reads from it.

How this is different from other Obsidian MCP servers. Existing Obsidian MCP servers (e.g. variants of mcp-obsidian) typically talk to the running Obsidian app through its Local-REST plugin and expose file-level tools — list_files, get_file, append_to_note, etc. second-brain-mcp is offline and plugin-free: it reads the vault as plain files from disk, builds a local semantic index with bge-m3, and exposes semantic search rather than path-based CRUD. Obsidian does not need to be running. The write path is a curated workflow with per-note human approval, not a raw write_file tool.

30-second quick start

# 1. Install (uvx — no venv pollution)
uvx second-brain-mcp serve --help

# 2. Point at your vault
export OBSIDIAN_VAULT=$HOME/obsidian/vault

# 3. Register with Claude Code
claude mcp add -s user second-brain \
  -e OBSIDIAN_VAULT="$OBSIDIAN_VAULT" \
  -- uvx second-brain-mcp serve

# 4. Restart Claude Code, then ask: "call obsidian_overview"

First run downloads the bge-m3 embedder (~2.3 GB) on the first tool call — roughly 5 seconds after that. See docs/CUSTOMIZE.md for lighter models.

Vault requirements

Hard minimum: a directory with .md files somewhere (any nesting).

Everything else is optional:

  • YAML frontmatter (type, verified, confidence) enables filtering
  • [[wikilinks]] enable backlinks navigation
  • _index.md at the vault root is returned whole in obsidian_overview

Start with your existing vault — unused features simply stay inactive until you add the relevant structure.

Feature No frontmatter No wikilinks No _index.md
semantic search
type_filter in search
obsidian_read body
obsidian_read frontmatter empty dict
obsidian_read outlinks/backlinks empty
obsidian_backlinks tool empty
obsidian_overview index_md placeholder

Why editorial, not archival

Most agent-memory systems default to the archival model: capture everything — raw conversation turns, every tool call, every message — then rely on semantic search to pull the right thing back later. Comparative retrieval tests against one such system (MemPalace, with its exchange-pair chunking and multi-layer palace) surfaced a clear trade-off: on realistic queries a small set of human-approved notes outperformed a much larger raw conversation archive sitting in the same index.

The reason is simple. In any given session roughly 95% of what's said is working noise — code, syntactic back-and-forth, tactical detail that expires with the task. The 5% that survives — atomic facts, decisions, insights — is what you actually want to find six months later. Archival memory keeps both and leans on the embedder to separate them, and that separation is hard to get right in practice.

second-brain-mcp takes the editorial position: memory is what you chose to remember. Nothing reaches the vault by accident. The loop:

  1. You work a session normally. The agent reads from the vault through the four read-only tools but writes nothing on its own.
  2. When you're done, you invoke to_obsidian. The agent walks the session, pulls out candidate facts, frames each as an atomic statement, and checks for duplicates against the existing vault.
  3. It shows you the list. You approve, reject, merge, or rewrite each candidate individually.
  4. Only the approved notes land in the vault.

The LLM does the extraction work — it is good at abstracting and generalising. You are the editor — you know which of the candidates actually matter. The vault ends up small, dense, and almost entirely signal. That curated remnant is what makes retrieval surface the right thing instead of the loudest thing.

This is why there is no reflection loop, no background extraction, no auto-writes. The approval gate is the whole point.

Documentation

License

MIT. See LICENSE.

Credits

The pattern of embedding a PROTOCOL string in the first read-tool's response is borrowed from MemPalace's tool_status. The design was shaped by empirical comparisons with raw conversation archives — curation beats volume.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

second_brain_mcp-1.0.1.tar.gz (32.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

second_brain_mcp-1.0.1-py3-none-any.whl (17.4 kB view details)

Uploaded Python 3

File details

Details for the file second_brain_mcp-1.0.1.tar.gz.

File metadata

  • Download URL: second_brain_mcp-1.0.1.tar.gz
  • Upload date:
  • Size: 32.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.12

File hashes

Hashes for second_brain_mcp-1.0.1.tar.gz
Algorithm Hash digest
SHA256 92b35059f9cabf2a749b96148e83a19b8cda9b763b0787e165a6e97664aa5693
MD5 cb843617f22c50a3b24b6a70852a18ce
BLAKE2b-256 6a1423bedd90776ce59f5a30b8f5b855933d42b4d3c00454c29fa7009d71df6d

See more details on using hashes here.

File details

Details for the file second_brain_mcp-1.0.1-py3-none-any.whl.

File metadata

File hashes

Hashes for second_brain_mcp-1.0.1-py3-none-any.whl
Algorithm Hash digest
SHA256 d355517fe42b6c11751b7f67515665f85b16e0f336f6e809bdb900e40d4fe432
MD5 ea8447d5a345817b2d3bb437c5fcbe67
BLAKE2b-256 95f1d9038ae80fdac5c0e2fd2ee0d5fc59c30b41bdb67188620b89424b9886eb

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page