Give your agent a knowledge graph that compounds.
Project description
Lacuna is a single MCP tool — wiki — that you plug into your existing agent harness (Claude Code, Hermes, OpenClaw) to give it a searchable, compounding personal research graph. Feed it a YouTube URL, an arXiv link, or a downloaded PDF. Your agent runs the structured extraction. The knowledge accumulates across every session.
Your vault is plain markdown — works natively with Obsidian. Browse your knowledge graph as a human, query it as an agent. Same files, no sync required.
It is the lacuna — the missing link between your raw inputs and your second brain.
Quick Start
pip install lacuna-wiki
lacuna init ~/my-vault
lacuna init creates your vault directory structure, sets up the DuckDB index in ~/.lacuna/, and asks whether to wire the MCP server into Claude Code and/or Hermes automatically. Takes about 10 seconds.
What Your Agent Unlocks
Once connected, your agent gets one composable tool:
wiki(q="attention mechanisms") # hybrid semantic + keyword search
wiki(page="transformer-architecture") # navigate to a specific page
wiki(pages=["sdpa", "flash-attn"]) # pull multiple pages in one shot
wiki(q="...", scope="sources") # search raw source chunks directly
# sweep — audit and queue
wiki(link_audit=True) # vault audit: research gaps, ghost pages, sweep queue
wiki(link_audit="slug") # single-page audit + top synthesis candidates
wiki(link_audit="slug", mark_swept=True, cluster={...}) # mark page swept; optionally queue a cluster
# synthesise — read and write synthesis clusters
wiki(synthesise=True) # list pending synthesis clusters
wiki(synthesise=N) # detail for cluster N: members, paths, coverage scores
wiki(synthesise=N, commit={"slug":"…"}) # mark cluster synthesised; links synthesis page in DB
That's it. One tool. Your entire research graph.
Omnivorous Inputs
Feed Lacuna anything — it knows what to do:
| Source | Command |
|---|---|
| 📺 YouTube URL | lacuna add-source https://youtube.com/watch?v=... |
| 📄 arXiv link | lacuna add-source https://arxiv.org/abs/2310.06825 |
| 📑 Local PDF | lacuna add-source ~/papers/my-paper.pdf |
| 🌐 Any URL | lacuna add-source https://example.com/blogpost |
The Structured Skills
This is where Lacuna is different from dropping a folder of PDFs into a vector store.
Lacuna ships with agent skills for Claude Code and Hermes that encode a structured, multi-turn extraction workflow — not "summarize this" but a disciplined process that produces tagged, wikilinked pages with full citations. When your agent ingests a paper, it follows the skill's protocol: pulling core concepts, mapping relationships to your existing graph, and flagging gaps.
Install them into your harness:
lacuna install-skills --claude-global # → ~/.claude/skills/
lacuna install-skills --hermes-global # → ~/.hermes/skills/
lacuna install-skills --openclaw-global # → ~/.openclaw/skills/
lacuna install-skills --hermes PATH # custom Hermes skills directory
Skills included:
- ingest — structured multi-turn knowledge extraction from a source
- query — cited, honest answers from your graph (flags what's missing)
- adversary — re-verifies old claims against their cited sources
- sweep — audits the vault for missing
[[wikilinks]], adds them, and queues related pages as synthesis candidates - synthesise — reads the synthesis queue and writes unified pages from clusters of related content
The Compounding Graph
Lacuna outputs aren't isolated notes. Each extraction is structured to deliberately compound — new pages wikilink to existing ones, concepts accumulate across sessions, and the graph gets richer with every source you add.
Under the hood: hybrid BM25 + vector search over a DuckDB store. No format lock-in — your vault is just a folder.
my-vault/
├── wiki/ # compiled knowledge pages (Obsidian-readable)
│ ├── attention.md
│ ├── transformer-architecture.md
│ └── ...
├── raw/ # original sources
│ ├── vaswani2017/
│ └── ...
└── .lacuna.toml # vault config
Embedding Backend
Lacuna needs an OpenAI-compatible embeddings endpoint. The easiest path is Ollama:
# Install Ollama: https://ollama.com/download
ollama pull nomic-embed-text:v1.5
Then set your vault's .lacuna.toml (created by lacuna init):
[embed]
url = "http://localhost:11434" # Ollama's default port
model = "nomic-embed-text:v1.5" # default — can omit
dim = 768 # default — can omit
[worker]
sync_workers = 4 # parallel threads for initial_sync (default: 4)
embed_concurrency = 4 # simultaneous embed requests (default: 4)
reader_pool_size = 3 # read connections for MCP + status API (default: 3)
LACUNA_EMBED_URL, LACUNA_EMBED_MODEL, LACUNA_EMBED_DIM, LACUNA_SYNC_WORKERS, LACUNA_EMBED_CONCURRENCY, and LACUNA_READER_POOL_SIZE env vars also work for one-off overrides.
Changing models? Set
embed.dimin.lacuna.tomlbefore runninglacuna init— the schema is created from that value. Changing the model or dim after ingesting sources will invalidate existing embeddings. Alacuna reindexcommand to re-embed everything in place is planned; for now, delete~/.lacuna/vaults/<your-vault>/and re-runlacuna initto start fresh.
Requirements
- Python 3.11+
pdftotext(poppler-utils) for PDF extraction:apt install poppler-utils/brew install poppler- An embedding server (Ollama, OpenAI, or any OpenAI-compatible endpoint)
Status
Early release. The core loop — add source → agent ingests → agent queries — is solid. The structured skills are where the value is; treat them as opinionated defaults you can adapt.
Windows support is in progress (Linux/macOS fully supported today).
Keeping the Graph Tidy
Ingest adds knowledge — sweep and synthesise maintain it.
Sweep audits the vault for missing [[wikilinks]] and detects pages that are converging on the same concept. For each page in the backlog, the agent reads it, adds any missing links one at a time, and declares a synthesis cluster if multiple pages are describing the same concept from different angles. Run it periodically in Claude Code:
/lacuna-sweep
After a large ingest, pre-warm the candidate cache before running the sweep skill so it doesn't time out on big vaults:
lacuna sweep # process all pages in the backlog
lacuna sweep --batch 50 # process the next 50 pages
lacuna sweep --force # recompute all pages regardless of last_swept
When the daemon is running, lacuna sweep submits the job to the daemon and polls for completion — the DB stays locked to one writer. When no daemon is running, it runs directly.
Synthesise consumes the synthesis queue populated by sweep. It reads each cluster, writes a unified synthesis page from the combined content of the member pages, and marks the members as synthesised. The synthesis page surfaces shared ground, disagreements, and source provenance in one place:
/lacuna-synthesise
Both skills support an auto mode for unattended runs — pass "auto" or "just run it" when invoking.
lacuna status shows the full queue state at a glance:
┏━━━━━━━━━━━━━━━━━━━━┳━━━━━━┓
┃ Table ┃ Rows ┃
┡━━━━━━━━━━━━━━━━━━━━╇━━━━━━┩
│ pages │ 106 │
│ research gaps │ 8 │ ← stub pages awaiting sources
│ ghost pages │ 5 │ ← slugs linked but not yet created
│ sweep backlog │ 23 │ ← pages needing a sweep pass
│ synthesis queue │ 12 │ ← clusters ready for synthesise
│ synthesised pages │ 4 │ ← members absorbed into a synthesis page
│ sections │ 464 │
│ sources │ 19 │
└────────────────────┴──────┘
Manual MCP Setup
lacuna init handles all of this automatically. If you need to wire things by hand:
Claude Code
The daemon serves the MCP tool over SSE on mcp_port (default 7654). Point Claude Code at it directly — this avoids spawning a second process that would conflict with the daemon's DB lock:
claude mcp add --transport sse --scope user lacuna http://127.0.0.1:7654/sse
The daemon must be running (lacuna start) for Claude Code to connect. If you need the tool available before the daemon is started, you can fall back to the stdio transport instead:
claude mcp add --scope user -e LACUNA_VAULT=/path/to/my-vault -- lacuna /full/path/to/lacuna mcp
Find the full path with which lacuna.
Hermes (~/.hermes/config.yaml)
mcp_servers:
lacuna:
command: lacuna
args: [mcp]
env:
LACUNA_VAULT: /path/to/my-vault
OpenClaw
openclaw mcp set lacuna '{"command":"lacuna","args":["mcp"],"env":{"LACUNA_VAULT":"/path/to/my-vault"}}'
Upgrading
pip install --upgrade lacuna-wiki
lacuna sync
lacuna sync applies any schema migrations automatically — safe to run on every upgrade. If the daemon is running, stop it first (lacuna stop) and restart after sync.
License
MIT © Markus Williams, 2026
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file lacuna_wiki-2.2.0.tar.gz.
File metadata
- Download URL: lacuna_wiki-2.2.0.tar.gz
- Upload date:
- Size: 9.9 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
c2ae960ff8ad394a94d5a06a3488adfaced7bcb415c0f1a17c450b786269178a
|
|
| MD5 |
f2aa62a3619c6e405df8f89c77d2f3a9
|
|
| BLAKE2b-256 |
0ca9f84cb5fbdeb9497ddc9032bd03c5393a2505cc039a4d88dcb6c786a8352d
|
File details
Details for the file lacuna_wiki-2.2.0-py3-none-any.whl.
File metadata
- Download URL: lacuna_wiki-2.2.0-py3-none-any.whl
- Upload date:
- Size: 91.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
c53fea5a07f16a547d378cd8ccf5fe1c5172a334c2a46217f15bed7a2197ad3f
|
|
| MD5 |
b26325ddb9136ab69a3cbe262fd5ea87
|
|
| BLAKE2b-256 |
49f4c6a7b5936e886c5659520f099c93e2aa7340738ddb1c9d1e55ab4194da14
|