Agentic repo-to-wiki: scan any repository into a knowledge store with wiki pages, diagrams, and grounded Q&A.
Project description
rekipedia
Your AI tech lead — always available, always up to date.
rekipedia scans any repository into a portable SQLite knowledge store and gives every developer on the team an LLM-powered tech lead they can ask anything: "How does the auth flow work?", "What's the fastest way to add a new API endpoint?", "What broke the payment service last week?"
No hallucinations, no guessing — every answer is grounded in your actual codebase.
Key features
- Agentic wiki orchestration:
PlannerAgentdesigns the wiki structure dynamically based on your repo - Page importance scoring: planner assigns each page an importance score (0–100); nav sidebar sorts by priority
- DeepWiki-style sections: pages grouped into logical sections (
getting-started,architecture,core-components, etc.) - Context slicing: each page only receives the data it needs (~40–60% token reduction vs fixed-layout approach)
- Hybrid RAG Q&A: FAISS-indexed code chunks + wiki pages give the LLM full codebase context when answering questions
- Embed provider choice:
--embed-provider openai|ollama|azure|...— any litellm-compatible embedding model - Wiki export: bundle to a single Markdown file, ZIP archive, or structured JSON (
rekipedia export) - Incremental updates: only re-processes changed files after the first scan
- Grounded Q&A: answers cite real file paths and line numbers — no hallucinations
Quick start
via npm / npx (no install required)
npx rekipedia init .
npx rekipedia scan .
via uv / uvx (no install required)
uvx rekipedia init .
uvx rekipedia scan .
Permanent install
# Core (scan + serve + ask)
pip install rekipedia
# or
uv tool install rekipedia
# With RAG support (semantic embed + search — needs faiss-cpu + numpy ~100MB)
pip install "rekipedia[rag]"
# Homebrew (Go single binary — no Python needed)
brew tap unrealandychan/tap
brew install rekipedia
Commands
| Command | Description |
|---|---|
rekipedia init [REPO] |
Scaffold .rekipedia/ with config.yml and update .gitignore |
rekipedia scan [REPO] |
Full analysis — extracts symbols, synthesises wiki pages, exports JSON |
rekipedia update [REPO] |
Incremental refresh — re-extracts only changed files, keeps the rest |
rekipedia ask [QUESTION] |
Interactive Q&A REPL — streaming answers, Ctrl+C to quit |
rekipedia serve [REPO] |
Start a local web UI to browse wiki pages and ask questions |
rekipedia embed [REPO] |
Build (or rebuild) the FAISS semantic search index for hybrid RAG Q&A |
rekipedia export [REPO] |
Bundle the wiki to a single file (--format md|zip|json) |
LLM configuration
After running rekipedia init, edit .rekipedia/config.yml:
version: 1
ignore:
- .git
- node_modules
- __pycache__
- .rekipedia
languages:
- python
- typescript
llm:
model: ollama/llama4 # any litellm model string
api_key: "" # or set REKIPEDIA_API_KEY env var
base_url: "" # for local / self-hosted endpoints
temperature: 0.2
Supported providers (via litellm)
| Provider | Example model string |
|---|---|
| Ollama (local, free) | ollama/llama4 |
| OpenAI | gpt-5.5 |
| Anthropic | claude-opus-4-6 |
| Google Gemini | gemini/gemini-3.0-pro |
| Any OpenAI-compatible | set base_url in config |
Runtime overrides (env vars)
export REKIPEDIA_MODEL=gpt-5.5
export REKIPEDIA_API_KEY=sk-...
export REKIPEDIA_BASE_URL=https://my-proxy/v1
Output
rekipedia scan writes everything to .rekipedia/ inside your repo:
.rekipedia/
├── config.yml # your settings (committed)
├── store.db # SQLite knowledge store (git-ignored)
├── scan_meta.json # last scan metadata (model, timestamp, file count)
├── wiki/ # generated Markdown pages (3–15 pages, dynamically planned)
│ ├── index.md
│ ├── architecture-overview.md
│ ├── repository-structure.md
│ └── ... (pages vary by repo)
├── rag/ # RAG index (git-ignored)
│ ├── index.faiss # FAISS flat L2 index
│ └── chunks.json # source code chunks + metadata
├── diagrams/ # Mermaid diagram files
│ ├── module-graph.md
│ └── class-hierarchy.md
└── exports/ # JSON exports
├── symbols.json
├── relationships.json
└── manifest.json # run summary + metadata + page importance scores
Dynamically generates 3–15 wiki pages based on repo complexity (powered by PlannerAgent).
The wiki structure is designed dynamically by PlannerAgent based on what's actually present in your repo:
| Section | Example pages | When generated |
|---|---|---|
| Getting Started | index, installation, quick-start | Always |
| Architecture | architecture-overview, data-flow, repository-structure | ≥3 modules |
| Core Components | One page per major module | ≥2 modules |
| API Reference | cli-reference, python-api, rest-api | CLI/HTTP handlers found |
| Development | testing, contributing, ci-cd | Test files found |
| Ecosystem | integrations, deployment | ≥3 external deps |
Scan options
# Use a specific LLM model
rekipedia scan . --model gpt-5.5
# Skip Docker (run extractors in-process)
rekipedia scan . --no-docker
# Write output to a custom directory
rekipedia scan . --output-dir /tmp/wiki-output
# Enable debug logging (litellm, HTTP, full tracebacks)
rekipedia scan . --verbose
# Auto-embed for RAG after scan
rekipedia scan . --embed-model text-embedding-3-small --embed-provider openai
RAG / semantic search
rekipedia ask uses hybrid retrieval — wiki pages + FAISS-indexed code chunks — to answer questions with full codebase context.
# Build or rebuild the FAISS index
rekipedia embed .
# Custom embedding model + provider
rekipedia embed . --model text-embedding-3-small --provider openai
rekipedia embed . --model nomic-embed-text --provider ollama
# If your embed provider uses a DIFFERENT API key from your main LLM:
rekipedia embed . --model text-embedding-3-small --provider openai
# set embed_api_key in config.yml, or:
export REKIPEDIA_EMBED_API_KEY=sk-your-openai-key
# Or configure everything in .rekipedia/config.yml:
# llm:
# model: ollama/llama4 # main LLM (local)
# embed_model: text-embedding-3-small
# embed_provider: openai
# embed_api_key: sk-xxx # separate key for embed provider
# embed_base_url: "" # optional: custom endpoint
# Env var overrides (all optional):
export REKIPEDIA_EMBED_MODEL=nomic-embed-text
export REKIPEDIA_EMBED_PROVIDER=ollama
export REKIPEDIA_EMBED_API_KEY=sk-xxx
export REKIPEDIA_EMBED_BASE_URL=https://my-proxy.example.com/v1
The FAISS index is saved to .rekipedia/rag/index.faiss and chunked source code to .rekipedia/rag/chunks.json.
Export the wiki
# Single combined Markdown file (default)
rekipedia export . --format md --output ./wiki-export.md
# ZIP archive (one .md per page + manifest.json)
rekipedia export . --format zip --output ./wiki.zip
# Structured JSON (all pages + metadata)
rekipedia export . --format json --output ./wiki.json
Incremental update
After the first scan, rekipedia update only re-processes files whose SHA-256 has changed. Unchanged symbols and relationships are carried forward from the previous run — the wiki is refreshed in seconds.
rekipedia update . # auto-detect changed files
rekipedia update . --no-docker # skip Docker
If no previous scan is found, update automatically falls back to a full scan.
Ask the wiki
# Start interactive Q&A session (streams answers, Ctrl+C to quit)
rekipedia ask
rekipedia ask --repo ./my-project
rekipedia ask --model gpt-4o
# Single-shot mode (backward compat)
rekipedia ask -q "How does the auth flow work?"
Answers are grounded entirely in your wiki pages and symbol index — the LLM cannot hallucinate details that aren't in the scanned knowledge store. Answers are streamed token-by-token with a spinner while waiting.
Not happy with a generated page? See docs/customizing.md — you can pin pages, override prompts, change the writing style, or add your own pages that scans will never touch.
Serve the wiki
rekipedia serve . # opens browser at http://127.0.0.1:7070
rekipedia serve . --port 8080 # custom port
rekipedia serve . --no-browser # don't auto-open browser
- Browse generated wiki pages in a dark-themed web UI
- Ask questions with the same grounded Q&A (answers streamed via the web)
- Q&A history stored in SQLite
Prerequisites
- Python ≥ 3.11 (or
uvwhich manages its own Python) - Docker — optional; used for isolated extraction. Falls back to in-process runner automatically if Docker is not available (
--no-dockerforces in-process mode)
Using rekipedia with AI coding agents
rekipedia ships a Hermes agent skill (rekipedia-agent-skill.md) that teaches AI assistants (Copilot, Claude Code, Codex) to use rekipedia as their codebase intelligence layer:
- Copy
rekipedia-agent-skill.mdinto your Hermes skills directory - Any agent with the skill loaded will automatically scan + query rekipedia before diving into source files
- Dramatically reduces context window usage for large codebases
Development
# Install all deps
make dev
# Run tests
make test
# Lint
make lint
# Build wheel + npm tarball
make build
Release
PYPI_TOKEN=*** NPM_TOKEN=*** make release
# Full release: build + tag + push + PyPI + npm
make release-all PYPI_TOKEN=*** NPM_TOKEN=***
# With version bump
make release-all PYPI_TOKEN=*** NPM_TOKEN=*** VERSION=0.5.0
License
MIT — see LICENSE.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file rekipedia-0.9.10.tar.gz.
File metadata
- Download URL: rekipedia-0.9.10.tar.gz
- Upload date:
- Size: 405.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.13
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
96c174d4d0c44e772a7c7807c7235aaa4619c54ac80e45bf4d2206186aeb9689
|
|
| MD5 |
3c3a3be1153a6c8dd6463f310f099405
|
|
| BLAKE2b-256 |
a21873b5f10b366c595fda5887f78ecc8f85b4e54601520c014f43dd96bce2c1
|
File details
Details for the file rekipedia-0.9.10-py3-none-any.whl.
File metadata
- Download URL: rekipedia-0.9.10-py3-none-any.whl
- Upload date:
- Size: 88.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.13
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
27e8d7997b97eadbbc3215503acec99066af806de52f1fedd33b314c626ae6ce
|
|
| MD5 |
9e47927ad3b9cb426be13b4e375f6858
|
|
| BLAKE2b-256 |
d9c0d501b439b83238816badd3d4854fe75f2b5a24979d8285c2f2b3f17f2ae3
|