Personal knowledge base that runs inside any MCP-compatible AI tool — structured, navigable, zero extra cost
Project description
kvault
Tell your AI agent to build you a knowledge base. That's it.
pip install knowledgevault[mcp]
kvault gives your coding agent persistent, structured memory. It runs as an MCP server inside Claude Code (or any MCP-compatible tool), using the subscription you already pay for. No extra API keys. No extra cost.
Your agent creates entities (people, projects, notes), navigates the hierarchy via parent summaries, and keeps everything in sync — all through 15 MCP tools.
Who is this for?
Developers using Claude Code, OpenAI Codex, Cursor, VS Code + Copilot, or any MCP-compatible tool who want their agent to remember things between sessions — contacts, projects, meeting notes, research — in a structured, navigable format.
What makes it different?
| kvault | Anthropic memory server | Notion AI / Mem.ai | obsidian-claude-pkm | |
|---|---|---|---|---|
| Structure | Hierarchical entities with navigable tree | Flat JSON | Rich docs, flat search | Obsidian vault |
| Agent-native | 15 MCP tools, built for agents | 4 tools, basic | Chat sidebar | Template, not runtime |
| Cost | $0 (uses existing subscription) | $0 | $12-20/mo extra | $0 |
| Navigation | Parent summaries at every level | None | AI-generated | Manual |
| Search | Agent uses its own Grep/Glob/Read | Built-in | Built-in | Manual |
Quickstart (30 seconds)
1. Install
pip install knowledgevault[mcp]
2. Add the MCP server to your AI tool
Claude Code (.claude/settings.json):
{
"mcpServers": {
"kvault": { "command": "kvault-mcp" }
}
}
OpenAI Codex (.codex/config.toml):
[mcp_servers.kvault]
command = "kvault-mcp"
Cursor (.cursor/mcp.json):
{
"mcpServers": {
"kvault": { "command": "kvault-mcp" }
}
}
VS Code + GitHub Copilot (.vscode/mcp.json):
{
"servers": {
"kvault": { "command": "kvault-mcp", "type": "stdio" }
}
}
Windsurf (~/.codeium/windsurf/mcp_config.json):
{
"mcpServers": {
"kvault": { "command": "kvault-mcp" }
}
}
3. Tell your agent
"Create a knowledge base for me at ./my_kb"
Your agent calls kvault_init, creates the directory structure, and you're up.
Try it: import your ChatGPT history
The best way to see kvault in action is to point it at data you already have. ChatGPT lets you export your entire conversation history — years of questions, people mentioned, projects discussed, decisions made — and Claude Code + kvault can turn it into a structured, navigable knowledge base in minutes.
1. Export your ChatGPT data
Go to ChatGPT → Settings → Data controls → Export data. You'll get an email with a zip file containing conversations.json.
2. Unzip it into your KB
unzip chatgpt-export.zip -d my_kb/sources/chatgpt
3. Tell Claude Code to process it
Read through my ChatGPT export in sources/chatgpt/conversations.json.
Extract the people, projects, and ideas I've discussed most frequently.
Create entities for each one in the knowledge base.
Claude Code will use the kvault tools to create structured entries with frontmatter and propagate summaries. You'll end up with a browsable, navigable knowledge base built from years of conversations you've already had.
Other great data sources to try:
| Source | How to get it | What you'll extract |
|---|---|---|
| ChatGPT history | Settings → Export data | People, projects, decisions, research threads |
| Google Contacts | Google Takeout (Contacts) | Names, emails, phone numbers, notes |
| iMessage | ~/Library/Messages/chat.db (macOS) |
Relationships, interaction frequency, context |
| Gmail | Google Takeout (Mail) | Professional contacts, threads, follow-ups |
| Meeting notes | Any folder of markdown/text files | People, action items, decisions |
| Notion export | Notion → Settings → Export | Projects, notes, wikis |
The pattern is always the same: drop the data into sources/, tell your agent to process it, and let kvault handle structure and propagation.
What happens next
Every time your agent processes new information, it follows a 4-step workflow:
- Navigate — Browse the tree, read parent summaries to understand what exists
- Write — Create/update entity with YAML frontmatter (
_summary.md) - Propagate — Update all ancestor
_summary.mdfiles so summaries stay in sync - Log — Add entry to
journal/YYYY-MM/log.md
Your agent uses its own Grep/Glob/Read tools for searching. Parent summaries at each level act as curated indexes — when reading any entity, kvault automatically includes the parent summary so the agent sees sibling context for free.
What an entity looks like
Each entity is a directory with a single _summary.md file containing YAML frontmatter:
---
created: 2026-02-06
updated: 2026-02-06
source: manual
aliases: [Sarah Chen, sarah@anthropic.com]
email: sarah@anthropic.com
relationship_type: colleague
---
# Sarah Chen
Research scientist at Anthropic working on causal discovery.
## Background
Met at NeurIPS 2025. Collaborator on interpretability project.
## Interactions
- 2026-02-06: Coffee meeting — discussed causal representation learning
## Follow-ups
- [ ] Share CJE paper draft
Required frontmatter: source, aliases (the MCP tools set created/updated automatically)
What a knowledge base looks like
my_kb/
├── _summary.md # Root: executive overview
├── people/
│ ├── _summary.md # "12 contacts across 3 categories"
│ ├── family/
│ │ ├── _summary.md
│ │ └── mom/
│ │ └── _summary.md
│ ├── friends/
│ │ ├── _summary.md
│ │ └── alex_rivera/
│ │ └── _summary.md
│ └── contacts/
│ ├── _summary.md
│ ├── sarah_chen/
│ │ └── _summary.md
│ └── james_park/
│ └── _summary.md
├── projects/
│ ├── _summary.md
│ └── cje_paper/
│ └── _summary.md
├── journal/
│ └── 2026-02/
│ └── log.md
└── .kvault/
└── logs.db # Observability
Every directory with a _summary.md is a node. Summaries at each level capture the semantic landscape of their children. When reading any entity, kvault returns the parent summary too — so the agent always knows what siblings exist.
MCP tools (15)
| Category | Tools |
|---|---|
| Init | kvault_init, kvault_status |
| Entity | kvault_read_entity, kvault_write_entity, kvault_list_entities, kvault_delete_entity, kvault_move_entity |
| Summary | kvault_read_summary, kvault_write_summary, kvault_get_parent_summaries, kvault_propagate_all |
| Workflow | kvault_log_phase, kvault_write_journal, kvault_validate_transition |
| Validation | kvault_validate_kb |
kvault_read_entity returns entity content plus the parent _summary.md — giving the agent sibling context for free. kvault_write_entity returns a propagation_needed list of ancestor paths, so agents know exactly which summaries to update.
Python API
kvault also exposes a Python API for programmatic use:
from pathlib import Path
from kvault import SimpleStorage, scan_entities
kg_root = Path("my_kb")
storage = SimpleStorage(kg_root)
# Scan all entities
entities = scan_entities(kg_root)
# Navigate hierarchy
ancestors = storage.get_ancestors("people/contacts/sarah_chen")
# Returns: ["people/contacts", "people"]
Integrity hook
Catch stale summaries before each prompt by adding to .claude/settings.json:
{
"hooks": {
"UserPromptSubmit": [
{
"type": "command",
"command": "kvault check --kb-root /absolute/path/to/my_kb"
}
]
}
}
Development
# Install dev dependencies
pip install -e ".[dev]"
# Run tests
pytest
# Lint, format, type-check
ruff check . && black . && mypy .
License
MIT
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file knowledgevault-0.6.0.tar.gz.
File metadata
- Download URL: knowledgevault-0.6.0.tar.gz
- Upload date:
- Size: 41.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
c81ca677284d5744853c41451f4c291c4e494ea36d636734263d6d746fa0d5bb
|
|
| MD5 |
0e0bb8229e365dce2518b0ee0d97e75b
|
|
| BLAKE2b-256 |
4a695526738ac813bb8513bbd88928ae5327b0ba5c659c96e1d5e5739450505d
|
Provenance
The following attestation bundles were made for knowledgevault-0.6.0.tar.gz:
Publisher:
publish.yml on cimo-labs/kvault
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
knowledgevault-0.6.0.tar.gz -
Subject digest:
c81ca677284d5744853c41451f4c291c4e494ea36d636734263d6d746fa0d5bb - Sigstore transparency entry: 952461395
- Sigstore integration time:
-
Permalink:
cimo-labs/kvault@9dc86612f2904b2fbea8dda8a38f1ba3600ce62e -
Branch / Tag:
refs/heads/main - Owner: https://github.com/cimo-labs
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@9dc86612f2904b2fbea8dda8a38f1ba3600ce62e -
Trigger Event:
workflow_dispatch
-
Statement type:
File details
Details for the file knowledgevault-0.6.0-py3-none-any.whl.
File metadata
- Download URL: knowledgevault-0.6.0-py3-none-any.whl
- Upload date:
- Size: 37.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
d6c702fdfb15137040513dc7d5a6cf4a78e63b7d1237c7489b9d74142e3e05be
|
|
| MD5 |
95b92b88392ece5bd5d48eadb75365aa
|
|
| BLAKE2b-256 |
638b769ef21e6656921fd9ef07ca2cd6c9b87aa7b916c754854795ffde710682
|
Provenance
The following attestation bundles were made for knowledgevault-0.6.0-py3-none-any.whl:
Publisher:
publish.yml on cimo-labs/kvault
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
knowledgevault-0.6.0-py3-none-any.whl -
Subject digest:
d6c702fdfb15137040513dc7d5a6cf4a78e63b7d1237c7489b9d74142e3e05be - Sigstore transparency entry: 952461398
- Sigstore integration time:
-
Permalink:
cimo-labs/kvault@9dc86612f2904b2fbea8dda8a38f1ba3600ce62e -
Branch / Tag:
refs/heads/main - Owner: https://github.com/cimo-labs
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@9dc86612f2904b2fbea8dda8a38f1ba3600ce62e -
Trigger Event:
workflow_dispatch
-
Statement type: