Autonomous data analyst: text-to-SQL with conversation, insight, monitoring, and community intelligence
Project description
TalkDB
An open-source, MCP-native autonomous data analyst. Converts natural-language questions into validated SQL, then goes further: multi-turn conversations, statistical insight generation, proactive metric monitoring, dual-path result verification, and a learning loop from user corrections.
Not just text-to-SQL — a full analyst loop in a single tool.
What it does
- Ask questions in plain English over Postgres, SQLite, and more. Answers include SQL, results, a narrative summary, and an auto-generated chart.
- Multi-turn conversations. "Revenue by month" → "just Q4" → "break that down by region" — follow-ups resolve to the previous turn's context.
- Semantic layer (YAML). Define business metrics once (
revenue = SUM(orders.total_amount) WHERE status='completed'); the LLM uses your definitions instead of guessing. - Dual-path verification. Every novel query is generated two structurally different ways; results are compared. Divergence drops confidence and surfaces a warning — catching semantic errors that schema validation can't.
- Confidence scoring with graceful refusal. Queries below threshold aren't silently wrong — they return a refusal explaining what's uncertain.
- Statistical insight agent. After results land, a pandas-based analyzer detects trends, anomalies, and concentrations. A chart is auto-generated from data shape. An LLM narrator writes 2–4 sentences using only the analyzer's facts (no hallucinated numbers).
- Proactive watchdog. Save any query as a scheduled watch. APScheduler runs it on your cadence, compares to a rolling baseline, and fires a webhook/Slack/stdout alert when conditions trigger.
- Self-improving via corrections.
correct_query(question, wrong_sql, correct_sql)stores the pattern and indexes it into retrieval, so future similar questions benefit.
Interfaces
- MCP server (primary). Works with Claude Desktop, Cursor, VS Code, and any MCP-compatible client. 12 tools:
ask,analyze,follow_up,list_databases,describe_database,validate_sql,correct_query,watch,list_watches,remove_watch,run_watch,get_session. - CLI.
talkdb ask,talkdb chat,talkdb init,talkdb index,talkdb watchdog add/list/remove/run/start,talkdb serve.
Tech stack
Python 3.11+ · FastMCP · LiteLLM (Claude / GPT / Gemini / Ollama) · SQLAlchemy 2 · sqlglot · ChromaDB + BM25 hybrid retrieval · pandas + matplotlib + seaborn · APScheduler · Pydantic.
Quick start
# Setup
python3 -m venv .venv && source .venv/bin/activate
pip install -e ".[dev]"
cp .env.example .env
# Edit .env: set OPENAI_API_KEY or ANTHROPIC_API_KEY, and TALKDB_DEFAULT_DB
# Seed a demo ecommerce DB (optional — or point at your own)
python scripts/seed_example_db.py
# Index schema + semantic model for retrieval
talkdb index
# Ask a question
talkdb ask "What is our total revenue?"
# Multi-turn chat
talkdb chat
# Start MCP server (for Claude Desktop / Cursor)
talkdb serve --transport stdio
Wiring into Claude Desktop
Add to ~/Library/Application Support/Claude/claude_desktop_config.json:
{
"mcpServers": {
"talkdb": {
"command": "/absolute/path/to/.venv/bin/talkdb",
"args": ["serve", "--transport", "stdio"],
"cwd": "/absolute/path/to/project"
}
}
}
Restart Claude Desktop. The ask, analyze, follow_up, etc. tools appear in the tool picker.
Semantic model
Define business meaning once in semantic_models/<db>.yaml:
metrics:
- name: revenue
description: "Total revenue from completed orders."
calculation: "SUM(orders.total_amount) WHERE orders.status = 'completed'"
table: orders
tables:
- name: customers
columns:
- name: tier
valid_values: ["bronze", "silver", "gold", "platinum"]
aliases: ["loyalty level", "membership tier"]
joins:
- left: orders
right: customers
on: "orders.customer_id = customers.id"
type: "INNER JOIN"
Run talkdb init --database mydb --output semantic_models/mydb.yaml to auto-generate a skeleton from your DB.
Watchdog example
talkdb watchdog add \
--name "Revenue monitor" \
--question "What is today's total revenue?" \
--schedule "every 1 hour" \
--alert "drops more than 20% below 7-day average" \
--slack-webhook "https://hooks.slack.com/services/..."
talkdb watchdog list
talkdb watchdog start # blocks; runs the scheduler
Alerts render like:
🔴 Revenue monitor — Current value: $38.2k — 27% below baseline $52.4k (7_day_avg). Suggested follow-up: "Why is today's revenue below baseline?"
Benchmark
Ships with a 27-case mini benchmark on the seeded DB for regression tracking:
python -m tests.benchmarks.run_benchmark
Current baseline: 23/27 (85%) execution accuracy, 96% structural correctness, 0 silent-wrong refusals. Per-phase baselines are checked into tests/benchmarks/.
Project layout
src/talkdb/
├── core/ # Engine, SQL generator, intent classifier, dialect transpiler
├── conversation/ # Session state, rewriter, reference resolver
├── schema/ # Introspector, data models, semantic model loader
├── retrieval/ # ChromaDB + BM25 hybrid retriever, embeddings
├── validation/ # Schema, execution, shape, dual-path, confidence
├── insight/ # Analyzer (pandas), charter (matplotlib), narrator (LLM)
├── watchdog/ # Scheduler, baseline, alerter, storage
├── learning/ # Pattern store, feedback recorder
├── connectors/ # Postgres, SQLite (more dialects via sqlglot)
├── server/ # FastMCP server
└── config/ # Pydantic settings
Design principles
- Never dump full schema into prompts. Hybrid retrieval (BM25 + vector) surfaces only relevant context.
- Never return results below confidence threshold. Refuse and explain — silent wrong answers destroy trust.
- SELECT only. sqlglot AST walk rejects every mutating statement before execution.
- Read-only validation. All validation queries run in
READ ONLYtransactions withLIMIT 10and a 10-second timeout. - LLM-provider agnostic. Every LLM call goes through LiteLLM — swap Claude ↔ GPT ↔ Gemini ↔ Ollama with one config change.
- Vector store abstraction. ChromaDB for dev, pgvector can plug in behind the same interface.
- Temperature 0 for SQL generation. Deterministic output.
- Insight stats are pandas, not LLM. Only narration uses the LLM (so the numbers in insights are always correct).
- Dual-path uses structurally different prompts. Path A direct, Path B decompose-then-compose. Catches correlated errors that self-correction misses.
- Conversation rewriting, not SQL mutation. Follow-ups are rewritten into standalone questions before SQL is regenerated from scratch.
License
MIT.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file talkdb_ai-0.1.0.tar.gz.
File metadata
- Download URL: talkdb_ai-0.1.0.tar.gz
- Upload date:
- Size: 94.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
77fc3550647b387f28967414afb48516dd994b317234ddc30e46d8a91ee29907
|
|
| MD5 |
dade843a0a4e8cbad3128676b0fc1a5c
|
|
| BLAKE2b-256 |
356ab3788e50a6992a84cc2b9f5561ac81244710bf1e5f2117aae863126096be
|
File details
Details for the file talkdb_ai-0.1.0-py3-none-any.whl.
File metadata
- Download URL: talkdb_ai-0.1.0-py3-none-any.whl
- Upload date:
- Size: 79.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
965708b397b40e781f00603cf295f82eefdd2e2a4a12a3c10605ddd329cd815e
|
|
| MD5 |
e9f7b133ae0a078c28670725a6765327
|
|
| BLAKE2b-256 |
cbc04a70c14d6d725c98d145a9717e55a9d8605db1b6ec093a8d07019a488f57
|