Skip to main content

Personal AI-powered knowledge base with RAG

Project description

stache-ai

A Python library for building AI-powered knowledge bases using Retrieval-Augmented Generation (RAG).

Overview

stache-ai provides a pluggable framework for ingesting documents, storing embeddings, and executing semantic search with optional reranking. It includes support for multiple vector databases, LLM providers, embedding models, and document formats.

Installation

Install the core package:

pip install stache-ai

Quick Start

from stache_ai.rag.pipeline import get_pipeline

# Get the pipeline (uses configured providers)
pipeline = get_pipeline()

# Ingest text
result = pipeline.ingest_text(
    text="Your knowledge base content here",
    metadata={"source": "example"}
)
print(f"Created {result['chunks_created']} chunks")

# Search
results = pipeline.query(
    question="What is this about?",
    top_k=5
)
for source in results['sources']:
    print(f"- {source['text'][:100]}...")

Provider Packages

stache-ai uses a provider pattern to support different backends. Install optional provider packages to enable specific functionality:

AWS Providers

pip install "stache-ai[aws]"

Includes:

  • stache-ai-s3vectors - Amazon S3 Vectors for semantic search
  • stache-ai-dynamodb - Amazon DynamoDB for namespace and document index storage
  • stache-ai-bedrock - Amazon Bedrock for LLMs and embeddings

Ollama

pip install "stache-ai[ollama]"

Includes:

  • stache-ai-ollama - Ollama for local LLM and embedding models

OpenAI

pip install "stache-ai[openai]"

Includes:

  • stache-ai-openai - OpenAI for GPT models and embeddings

Configuration

Configure stache-ai via environment variables or a .env file:

# Vector Database
VECTORDB_PROVIDER=s3vectors
VECTORDB_S3_REGION=us-east-1
VECTORDB_S3_INDEX_NAME=stache

# Embeddings
EMBEDDING_PROVIDER=bedrock
EMBEDDING_MODEL=cohere.embed-english-v3

# Namespaces
NAMESPACE_PROVIDER=dynamodb
NAMESPACE_DYNAMODB_TABLE=stache-namespaces

# LLM
LLM_PROVIDER=bedrock
LLM_MODEL=anthropic.claude-3-5-sonnet-20241022-v2:0

# Optional features
ENABLE_DOCUMENT_INDEX=true
EMBEDDING_AUTO_SPLIT_ENABLED=true

See src/stache_ai/config.py for all available options.

Usage Examples

Document Chunking

from stache_ai.chunking import ChunkingStrategy

# Recursive character-level chunking
chunks = ChunkingStrategy.create(
    strategy="recursive",
    chunk_size=1024,
    chunk_overlap=100
).chunk("Your document text")

for chunk in chunks:
    print(chunk)

Filtering Results

# Search with metadata filter
results = pipeline.query(
    question="API documentation",
    filter={"source": "docs"}
)

Namespace Isolation

# Ingest to a specific namespace
pipeline.ingest_text(
    text="Project A data",
    namespace="project-a"
)

# Search within a namespace
results = pipeline.query(
    question="Find related content",
    namespace="project-a"
)

API Server

Run a FastAPI server for HTTP access:

pip install stache-ai[dev]
python -m stache_ai.api.main

Server exposes endpoints for:

  • /api/query - Semantic search
  • /api/capture - Text ingestion
  • /api/namespaces - Manage namespaces
  • /api/documents - List and retrieve documents
  • /api/upload - Upload files (PDF, DOCX, etc.)

CLI Tools

Admin CLI (stache-admin)

# Import documents from a directory
stache-import /path/to/documents --namespace my-docs

# List namespaces
stache-admin namespace-list

# View vector statistics
stache-admin vectors stats

User CLI (stache-tools)

For search, ingest, and MCP server, install stache-tools:

pip install stache-tools

# Search
stache search "your query"

# Ingest text
stache ingest -t "your text" -n namespace

Testing

pip install stache-ai[dev]
pytest

Documentation

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

stache_ai-0.1.3.tar.gz (120.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

stache_ai-0.1.3-py3-none-any.whl (130.5 kB view details)

Uploaded Python 3

File details

Details for the file stache_ai-0.1.3.tar.gz.

File metadata

  • Download URL: stache_ai-0.1.3.tar.gz
  • Upload date:
  • Size: 120.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.3

File hashes

Hashes for stache_ai-0.1.3.tar.gz
Algorithm Hash digest
SHA256 50bf6b144b991101126113421f113ac7ca355042b292b2fcd9082bbb1842aa1f
MD5 3c8eeb1c056c094a4fe93de092ed5d13
BLAKE2b-256 f2021ea21e3bc01051f77aa77f65506640f3313b90c5fefee8446d6dd2d480af

See more details on using hashes here.

File details

Details for the file stache_ai-0.1.3-py3-none-any.whl.

File metadata

  • Download URL: stache_ai-0.1.3-py3-none-any.whl
  • Upload date:
  • Size: 130.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.3

File hashes

Hashes for stache_ai-0.1.3-py3-none-any.whl
Algorithm Hash digest
SHA256 f7ad3b60ccc38ca5f6c1eb193d647b6a1de790df213584347655a08719f39d66
MD5 8d731bf50b1f9ce3845feaf1121fa121
BLAKE2b-256 c8a7b6403ab7375de5f768ce31f946089a62bea5fda65a058a8923183581b0f0

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page