Skip to main content

Turn your AI conversations into a searchable second brain with cognitive prosthetic tools

Project description

🧠 brain-mcp

You've had thousands of AI conversations. You can't search any of them.

Docs Python 3.11+ MCP LanceDB License PyPI CI

brain-mcp Demo

⬆️ Auto-playing preview — click here for full video with audio

📚 Documentation · 🚀 Quickstart · ❓ FAQ · 🔧 All 25 Tools

If this is useful, a ⭐ helps others find it.


Install

pipx install brain-mcp        # recommended (isolated env, on your PATH)
brain-mcp init --full          # discover conversations, import, embed
brain-mcp setup claude         # auto-configure Claude Desktop + Code

Restart Claude. Done. 25 tools available.

Alternative: pip install
pip install brain-mcp
brain-mcp init --full
brain-mcp setup claude

Note: If you install in a virtualenv, make sure brain-mcp is on your PATH — Claude Desktop/Code needs to find the binary. pipx handles this automatically.

Setup for specific clients
brain-mcp setup claude           # auto-detect: configures both Desktop + Code
brain-mcp setup claude-desktop   # Claude Desktop only
brain-mcp setup claude-code      # Claude Code only
brain-mcp setup cursor           # Cursor
brain-mcp setup windsurf         # Windsurf

What You Can Do

Ask your AI What happens
"What did I figure out about sleep last month?" Finds your insights across 12 conversations you forgot you had
"Search everything I've discussed about marketing" 23 conversations across 8 months, with quotes — in 12ms
"Where did I leave off with the business plan?" Reconstructs your context — open questions, decisions, next steps
"How has my thinking about career changes evolved?" Tracks your opinion trajectory from doubt → clarity
"What would it cost to switch focus right now?" Quantifies what you'd abandon — open threads, unfinished thinking
"What do I actually think about AI?" Synthesizes YOUR views from 31 past conversations into one answer

Works for researchers, writers, students, founders, developers — anyone who thinks with AI.


The Problem Nobody Talks About

You had a breakthrough at 2am last Tuesday. You laid out a whole framework in a conversation with Claude. It was brilliant.

You can't find it. You can't even remember which conversation it was in.

Every week, millions of people pour their best thinking into AI conversations — and lose all of it. ChatGPT's "memory" stores a few fun facts. Claude's import tool gives you a markdown summary of a summary. None of them let you search your own thinking.

brain-mcp doesn't store facts. It reconstructs cognitive state — where you were in a problem, what you'd decided, what questions were still open, and what it would cost to switch away.

Built with ADHD in mind. If your brain drops context constantly, this is your external hard drive.

Without vs. With

Without brain-mcp:

"I had this great idea about the business plan last month... let me search my chat history... which conversation was it... was it ChatGPT or Claude..."

30 minutes later: Maybe 60% recovered. If you're lucky.

With brain-mcp:

> "Where did I leave off with the business strategy?"

🧠 business-strategy — exploring stage
Open questions: 12 | Decisions made: 8

❓ Top open:
  - Should I focus on B2B or B2C first?
  - What pricing model fits the early stage?

✅ Recent decisions:
  - Target solo developers initially
  - Open-source core, paid hosting layer

💬 Found across: 15 ChatGPT + 8 Claude + 3 Claude Code conversations

⏱️ 12ms

12 milliseconds to reconstruct the mental state that took weeks to build. Across every AI tool you've used. That's real data, not a mockup.


How It Works

graph TD
    subgraph SOURCES["📥 Your Conversations (already on your machine)"]
        CC["Claude Code sessions"]
        CG["ChatGPT exports"]
        CB["Clawdbot sessions"]
        GJ["Generic JSONL"]
    end

    INIT["brain-mcp init --full"]

    subgraph DB["💾 Local Storage"]
        DDB["DuckDB (SQL)<br/>Full-text search"]
        LDB["LanceDB (768d vectors)<br/>Semantic search"]
        SUM["Structured summaries"]
    end

    TOOLS["🧠 25 MCP Tools"]

    subgraph CLIENTS["💬 Your AI Assistant"]
        CLA["Claude"]
        CUR["Cursor"]
        WS["Windsurf"]
    end

    SOURCES --> INIT --> DB --> TOOLS --> CLIENTS

    style SOURCES fill:#f8f9fa,stroke:#d0d7de,color:#1a1a1a
    style DB fill:#fff8e1,stroke:#d29922,color:#1a1a1a
    style CLIENTS fill:#f0fff4,stroke:#3fb950,color:#1a1a1a

All data stays on your machine. Embedding model runs locally (nomic-v1.5 on Apple Silicon). No cloud. No API costs for core operations.


25 Tools

🧠 Cognitive Prosthetic (8)

The tools that make this different from every other memory system.

Tool What it does Speed
tunnel_state "Load your save game" — reconstructs where you were in any domain 12ms
context_recovery Full re-entry brief with summaries, open questions, decisions 12ms
switching_cost Quantified cost of switching between domains 9ms
open_threads Everything unfinished, everywhere 2.7s
dormant_contexts Abandoned domains with open questions you forgot about 2.7s
cognitive_patterns When and how you think best, with data 10ms
tunnel_history Engagement timeline for a domain 5ms
trust_dashboard System-wide proof the safety net works 59ms

🔍 Search (6)

Tool What it does
semantic_search Vector search via LanceDB (768d nomic embeddings)
search_conversations Keyword search across all conversations
unified_search Search conversations + GitHub + markdown at once
search_summaries Structured summaries (extract: decisions/questions/quotes)
search_docs Markdown corpus search
unfinished_threads Threads with open questions by domain

🔬 Synthesis (4)

Tool What it does
what_do_i_think Synthesized view of your position on any topic
alignment_check Check decisions against your own stated principles
thinking_trajectory How an idea evolved over time
what_was_i_thinking Month-level snapshot of your focus

💬 Conversation + Stats (5)

get_conversation · conversations_by_date · brain_stats · query_analytics · github_search

⚙️ Meta (2)

list_principles · get_principle


Progressive Tiers

Every tool works at every tier — just with increasing depth:

What you have What works
Just conversations Keyword search, date browsing, stats
+ Embeddings Semantic search, synthesis, trajectory
+ Summaries Full prosthetic tools with structured domain analysis

Comparison

brain-mcp Mem0 Khoj Letta (MemGPT)
Memory model Conversation archaeology — reconstructs cognitive state Key-value fact store Hybrid search over docs Tiered agent memory
State recovery 8 prosthetic tools (tunnel state, switching cost, dormancy)
Data source Your existing AI conversations (auto-discovered) Runtime extractions Personal documents Agent conversation history
Runs where 100% local (Apple Silicon optimized) Cloud API or self-hosted Self-hosted or cloud Self-hosted or cloud
Domain tracking 25 cognitive domains with stages, open questions, decisions
Cost ~$0.05/day Free tier / paid Free / self-hosted Free / self-hosted
Protocol MCP (Claude, Cursor, any client) REST API REST API + web UI REST API

🔒 Privacy & Security

  • 100% local — all data stays on your machine
  • No telemetry — zero tracking, zero phone-home
  • No cloud dependency — works offline after initial setup
  • No accounts — no sign-up, no API keys for core features
  • You own everything — MIT licensed, your data is yours
  • Open source — audit every line of code

CLI

brain-mcp init              # Discover conversation sources
brain-mcp init --full       # Discover + import + embed (one command)
brain-mcp setup claude      # Configure Claude Desktop / Claude Code
brain-mcp setup cursor      # Configure Cursor
brain-mcp doctor            # Health check
brain-mcp sync              # Incremental update
brain-mcp status            # One-line status

Supported Sources

Source Auto-detected Status
Claude Code Supported
Claude Desktop Supported
ChatGPT Supported
Clawdbot Supported
Cursor Coming soon
Windsurf Coming soon
Generic JSONL Manual Supported

Requirements

  • Python 3.11+
  • ~500MB disk, ~2GB RAM for embedding
  • macOS (Apple Silicon recommended), Linux, or WSL

Part of the Ecosystem

Repo What
brain-mcp Memory — 25 MCP tools, cognitive prosthetic
QinBot AI on a $50 dumb phone — no browser, no apps
local-voice-ai Voice — Kokoro TTS + Parakeet STT, zero cloud
agent-memory-loop Cron + memory cascade for AI agents
brain-canvas Visual display for any LLM
x-search Search X/Twitter from terminal via Grok
mordenews Automated daily AI podcast
live-translate Real-time Hebrew→English translation

--

Contributing

See CONTRIBUTING.md for development setup, testing, and PR guidelines. All contributions welcome.


License

MIT — see LICENSE. See CHANGELOG.md for version history.


Built because losing your train of thought shouldn't mean starting over.

brainmcp.dev · Full Docs · PyPI

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

brain_mcp-0.1.3.tar.gz (71.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

brain_mcp-0.1.3-py3-none-any.whl (75.1 kB view details)

Uploaded Python 3

File details

Details for the file brain_mcp-0.1.3.tar.gz.

File metadata

  • Download URL: brain_mcp-0.1.3.tar.gz
  • Upload date:
  • Size: 71.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.3

File hashes

Hashes for brain_mcp-0.1.3.tar.gz
Algorithm Hash digest
SHA256 1f3e91ce4ca531f3cf62fd8120b43c08736f6d064c2a02a52a8485e203f9286c
MD5 bdf87ad5d5e4d2519438079e2ec6d867
BLAKE2b-256 5446dce81a30d9ea33936a978e158d3d8b7ebc5464f900542d4502000c462960

See more details on using hashes here.

File details

Details for the file brain_mcp-0.1.3-py3-none-any.whl.

File metadata

  • Download URL: brain_mcp-0.1.3-py3-none-any.whl
  • Upload date:
  • Size: 75.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.3

File hashes

Hashes for brain_mcp-0.1.3-py3-none-any.whl
Algorithm Hash digest
SHA256 94b84bd88a321b91df2c04d465cd0a1ea76766e1969029e385dcc7f8dbbef8a1
MD5 8221f58faee0764c75368761f509dea4
BLAKE2b-256 1870cf3b6ae7991687f1d6d86f0bf85501556a8134b3562c666ffdd3dbabfb5f

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page