Skip to main content

Compressed vector and graph-augmented retrieval engine with adaptive two-stage search, implementing TurboQuant/QJL quantization from Google Research.

Project description

TurboRAG

TurboRAG is a production-grade compressed vector retrieval engine with graph-augmented search, implementing the quantization techniques from Google Research's TurboQuant (ICLR 2026), Quantized Johnson-Lindenstrauss (QJL) (AAAI 2025), and PolarQuant (AISTATS 2026) papers by Amir Zandieh and Vahab Mirrokni.

Performance

Small Scale (1K vectors, 128-dim, 100 queries, k=10, 4-bit)

Backend Recall@10 MRR QPS Memory
TurboRAG 4-bit 1.000 1.000 6,209 0.08 MB
Exact float32 1.000 1.000 26,774 0.49 MB
FAISS Flat 1.000 1.000 32,384 0.49 MB
FAISS HNSW 1.000 1.000 23,640 0.55 MB
FAISS IVF-PQ 0.990 0.990 27,438 < 0.49 MB

Large Scale (100K vectors, 384-dim, 200 queries, k=10, 3-bit)

Latest local reproducible runs on arm64 as of 2026-04-12. Exact mode now uses the native threaded 3-bit top-k path with a 12-bit half-group fused scorer by default.

Backend Recall@10 MRR QPS Memory Notes
TurboRAG exact 1.000 1.000 66 18.3 MB Full side-by-side FAISS matrix run
TurboRAG fast 0.975 0.975 131 18.3 MB Binary sketch head + LUT refine
Exact float32 1.000 1.000 73 146.5 MB Brute force
FAISS Flat 1.000 1.000 60 146.5 MB Brute force
FAISS HNSW 0.610 0.610 771 152.6 MB High throughput, recall drop

Exact-only benchmark note: repeated TurboRAG exact-only runs on the same fixture now median at 70.98 QPS in the main benchmark environment.

TurboRAG auto mode selects the best strategy per query: exact for small indexes, fast for large ones. Exact mode uses the native C scorer and defaults to up to 8 threads. Override with TURBORAG_EXACT_THREADS=<n> when benchmarking.

Key advantages:

  • 8× memory compression — 18.3 MB vs 146.5 MB float32 at 100K scale, with perfect recall in exact mode
  • Adaptive two-stage search — binary sketch pre-filter (SimHash + POPCNT) with full LUT refine gives about 2× throughput over current exact on the 100K fixture while holding 97.5% recall
  • Guaranteed exact modemode="exact" always gives perfect recall when accuracy is non-negotiable
  • Memory-first exact retrieval — exact mode keeps perfect recall at 18.3 MB instead of a 146.5 MB float32 matrix
  • Production-hardened with atomic persistence, concurrency-safe HTTP service, and input validation

What Ships Today

  • Core engine: Compressed vector index with adaptive two-stage search (binary sketch head + fused LUT refine), C scoring kernel with POPCNT, batch search, threaded shard scanning
  • Graph retrieval: Entity extraction, community detection, hybrid dense+graph search with explainability
  • Document ingestion: Token-aware chunking for PDF, markdown, and plain text with metadata propagation
  • Sidecar adoption: Drop-in compatibility adapters for existing RAG systems — no database migration required
  • HTTP service: Production-hardened REST API with CORS, metrics, request tracking, batch queries, text ingestion, concurrency-safe mutations
  • Atomic persistence: save() clears stale shard files before writing, preventing deleted vectors from reappearing
  • MCP server: Tool-based agent integration over stdio (query, describe, ingest)
  • Benchmark suite: Side-by-side comparison against exact float and FAISS backends
  • CLI: Full command-line interface for import, query, benchmark, serve, and MCP modes
  • Client SDKs: TypeScript/Node.js, Go, and Ruby clients for language-agnostic HTTP integration
  • Docker: Production-ready multi-stage Dockerfile with pre-compiled C kernel

Package Layout

src/turborag/
  __init__.py
  _cscore.c            # C scoring kernel (auto-compiled)
  _cscore_wrapper.py   # ctypes bridge with auto-compilation
  adapters/            # LangChain-style and generic compatibility adapters
  benchmark.py         # Side-by-side benchmark harness
  chunker.py           # Token-aware PDF/MD/text chunking
  cli.py               # Click-based CLI with global logging options
  compress.py          # Rotation, quantization, LUT-based scoring
  embeddings.py        # Optional sentence-transformers integration
  exceptions.py        # Domain-specific exception hierarchy
  fast_kernels.py      # Vectorised LUT scoring (Python + C dispatch)
  graph.py             # Entity graph with persistence and community detection
  hybrid.py            # Dense + graph hybrid retrieval
  index.py             # TurboIndex with search, batch, delete, update
  ingest.py            # Dataset import and sidecar builder
  mcp_server.py        # MCP stdio server (query, describe, ingest tools)
  service.py           # Starlette HTTP service with CORS, metrics, batch
  types.py             # ChunkRecord, RetrievalResult dataclasses
tests/                 # 104+ tests
Dockerfile             # Multi-stage production build

Quick Start

import numpy as np
from turborag import TurboIndex

rng = np.random.default_rng(42)
index = TurboIndex(dim=8, bits=3, seed=7)
vectors = rng.normal(size=(100, 8)).astype(np.float32)
ids = [f"chunk-{i}" for i in range(len(vectors))]
index.add(vectors, ids)

results = index.search(vectors[0], k=5)
for chunk_id, score in results:
    print(chunk_id, score)

Install

From PyPI

# Core
pip install turborag

# Everything
pip install turborag[all]

From Source

git clone https://github.com/ratnam1510/turborag.git
cd turborag
pip install -e '.[all,dev]'

Individual extras: graph, embed, ingest, serve, mcp, all, dev. See docs/installation.md for details.

Run As A Service

turborag serve --index ./my_index --host 0.0.0.0 --port 8080 --workers 4

Endpoints:

  • GET /health — health check
  • GET /index — index configuration and stats
  • GET /metrics — latency histograms and error counts
  • POST /query — single query (vector or text)
  • POST /query/batch — batch vector queries
  • POST /ingest — add records with embeddings
  • POST /ingest-text — raw text ingestion with auto-chunking

CORS is enabled by default. Use --cors-origins to restrict.

Docker

docker build -t turborag .
docker run -p 8080:8080 -v ./my_index:/data/index turborag \
  turborag serve --index /data/index --host 0.0.0.0

Side-By-Side Benchmark

# One-command local comparison
./scripts/benchmark_compare.sh

# Custom benchmark
turborag benchmark \
  --index ./turborag_sidecar \
  --queries ./queries.jsonl \
  --turborag-mode exact \
  --dataset ./corpus.jsonl \
  --baseline exact \
  --k 10

For the current exact-mode path, use TURBORAG_EXACT_THREADS=8 or leave it unset to use the default native thread count.

See docs/benchmarking.md.

MCP Agent Integration

turborag mcp --index ./my_index

Tools exposed: turborag_query, turborag_describe, turborag_ingest.

Sidecar Adoption (No Database Migration)

TurboRAG runs as a retrieval sidecar alongside your existing RAG database:

  1. Keep your current chunk/document store untouched
  2. Build a TurboRAG index from your existing embeddings
  3. Query TurboRAG for ranked IDs → hydrate from your existing DB
  4. Gradually shift traffic as confidence grows

Full guide: docs/current-rag-rollout.md.

Existing DB Integration (ID-Only, Low Memory)

If you already have text and metadata in your own database, TurboRAG can return ranked IDs only and skip local hydration entirely.

  • Python adapter: build from embeddings and pass your existing backend client.
  • HTTP service: use POST /query with "hydrate": false.
  • CLI: use turborag query --ids-only.

This mode minimizes memory and keeps TurboRAG as a pure retrieval sidecar.

from turborag.adapters.compat import ExistingRAGAdapter

adapter = ExistingRAGAdapter.from_existing_backend(
    embeddings=embeddings_matrix,
    ids=chunk_ids,
    query_embedder=embedder,
    records_backend=your_existing_db_client,
    bits=3,
)

hits = adapter.search_ids("What changed in capex guidance?", k=5)
# Hydrate hits from your own DB path.

Known backend helper builders are available in turborag.adapters.backends:

  • Postgres / Neon / Supabase Postgres: build_postgres_fetch_records(...) / build_neon_fetch_records(...)
  • Supabase Python client: build_supabase_fetch_records(...)
  • Pinecone: build_pinecone_fetch_records(...)
  • Qdrant: build_qdrant_fetch_records(...)
  • Chroma: build_chroma_fetch_records(...)

Or configure plug-and-play adapter mode via CLI:

turborag adapt set neon --index ./turborag_sidecar --option dsn=${DATABASE_URL}
turborag serve --index ./turborag_sidecar

Shortcut command style (env-aware):

turborag adapt supabase --index ./turborag_sidecar
turborag serve --index ./turborag_sidecar

Automatic mode (detect backend from env):

turborag adapt --index ./turborag_sidecar

If you're in the index directory already:

turborag adapt

Need a quick starter command for a backend?

turborag adapt demo supabase
turborag serve --index ./turborag_sidecar --no-load-snapshot

curl -X POST http://localhost:8080/query \
  -H 'content-type: application/json' \
  -d '{"query_vector":[0.1,0.2,0.3],"top_k":5,"hydrate":false}'

Verification

# Test suite (104+ tests)
pytest

# End-to-end smoke test
./scripts/smoke_test.sh

Documentation

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

turborag-0.4.2.tar.gz (96.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

turborag-0.4.2-py3-none-any.whl (78.7 kB view details)

Uploaded Python 3

File details

Details for the file turborag-0.4.2.tar.gz.

File metadata

  • Download URL: turborag-0.4.2.tar.gz
  • Upload date:
  • Size: 96.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.7

File hashes

Hashes for turborag-0.4.2.tar.gz
Algorithm Hash digest
SHA256 7f1c595017d132e9b4d15267344322c89588748c93c0f106860eba60a29d5529
MD5 4f2232327af9c091249f0c54922e8255
BLAKE2b-256 05ed30f61fb8cfd3e2ec7779eb3d12fcd0f0b113ff061d51056bef65d0648692

See more details on using hashes here.

File details

Details for the file turborag-0.4.2-py3-none-any.whl.

File metadata

  • Download URL: turborag-0.4.2-py3-none-any.whl
  • Upload date:
  • Size: 78.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.7

File hashes

Hashes for turborag-0.4.2-py3-none-any.whl
Algorithm Hash digest
SHA256 3398c81df7aeae209e45cc53369e3e4b1096cbd9da35764b5015c0ef2984097d
MD5 c6657d9afeb7774bd86f5a0ee3185cdf
BLAKE2b-256 f4c21f1b7a11f3c06953c8106e507a482ae05e37d99a3a08e00e1dd1abdbc467

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page