Skip to main content

Multi-Datalake RAG Indexer with Unified Collection & Per-Source Chunking - MCP & HTTP Server

Project description

Prometh Cortex

License PyPI version Python Support Tests Claude Code

Multi-Datalake RAG Indexer with Local MCP Integration

Table of Contents

Overview

Prometh Cortex is a local-first, extensible system for indexing multiple datalake repositories containing Markdown files and exposing their content for retrieval-augmented generation (RAG) workflows through a local MCP (Modular Command Processor) server.

Features

  • Multi-Datalake Support: Index multiple repositories of Markdown documents
  • YAML Frontmatter Parsing: Rich metadata extraction with structured schema support
  • Dual Vector Store Support: Choose between local FAISS or cloud-native Qdrant
  • Incremental Indexing: Smart change detection for efficient updates
  • MCP Server: Local server with stdio, SSE, and streamable HTTP transports for Claude, OpenCode, VSCode, and other tools
  • CLI Interface: Easy-to-use command line tools for indexing and querying
  • Performance Optimized: Target <100ms query response time on M1/M2 Mac

Installation

From PyPI (Recommended)

pip install prometh-cortex

From Source (Development)

git clone https://github.com/prometh-sh/prometh-cortex.git
cd prometh-cortex
pip install -e ".[dev]"

Quick Start

  1. Install via pip:
pip install prometh-cortex
  1. Initialize configuration (creates ~/.config/prometh-cortex/config.toml):
pcortex config --init
  1. Edit your config file:
# macOS/Linux
nano ~/.config/prometh-cortex/config.toml

# Or use your preferred editor
# Update the [datalake] repos with your document paths
  1. Build index:
pcortex build
  1. Query locally:
pcortex query "search for something"
  1. Start servers:
# For Claude Desktop (MCP protocol)
pcortex mcp

# For Perplexity/VSCode/HTTP integrations
pcortex serve

Configuration File Locations

Prometh-cortex follows the XDG Base Directory Specification. Config files are searched in this order:

  1. ./config.toml - Current directory (highest priority)
  2. ~/.config/prometh-cortex/config.toml - XDG config directory (recommended)
  3. ~/.prometh-cortex/config.toml - Fallback location

Useful commands:

# Initialize config in XDG directory (recommended for system-wide use)
pcortex config --init

# Create sample config in current directory (for project-specific config)
pcortex config --sample

# Show all config search paths
pcortex config --show-paths

Configuration

Create a config.toml file with your settings:

cp config.toml.sample config.toml
# Edit config.toml with your specific paths and settings

Configuration Format (TOML) - v0.5.0

Prometh Cortex v0.5.0 builds on the unified collection architecture with memory preservation during force rebuilds. A single FAISS/Qdrant index contains all documents plus an auto-injected virtual prmth_memory source for session summaries and decisions.

[datalake]
# Add your document directories here
repos = [
    "/path/to/your/notes",
    "/path/to/your/documents",
    "/path/to/your/projects"
]

[storage]
rag_index_dir = "/path/to/index/storage"

[server]
port = 8080
host = "localhost"
auth_token = "your-secure-token"
transport = "stdio"  # "stdio", "sse", or "streamable-http" (v0.4.0+)

[embedding]
model = "sentence-transformers/all-MiniLM-L6-v2"
max_query_results = 10

# Single unified collection
[[collections]]
name = "prometh_cortex"

# Multiple sources with per-source chunking parameters
[[sources]]
name = "knowledge_base"
chunk_size = 768
chunk_overlap = 76
source_patterns = ["docs/specs", "docs/prds"]

[[sources]]
name = "meetings"
chunk_size = 512
chunk_overlap = 51
source_patterns = ["meetings"]

[[sources]]
name = "todos"
chunk_size = 256
chunk_overlap = 26
source_patterns = ["todos", "reminders"]

[[sources]]
name = "default"
chunk_size = 512
chunk_overlap = 50
source_patterns = ["*"]  # Catch-all for unmatched documents

# Virtual memory source (v0.5.0+)
# Auto-injected; no file-based routing
# Stores session summaries, decisions, patterns
# Preserved during force rebuilds
[[sources]]
name = "prmth_memory"
chunk_size = 512
chunk_overlap = 50
source_patterns = [".prmth_memory"]  # Virtual pattern (won't match real files)

[vector_store]
type = "faiss"  # or "qdrant"

# Qdrant configuration (when type = "qdrant")
[vector_store.qdrant]
host = "localhost"
port = 6333
collection_name = "prometh_cortex"

Key Features in v0.5.0:

  • ✅ Memory preservation during pcortex build --force and pcortex rebuild
  • ✅ Virtual prmth_memory source auto-injected into all configs
  • ✅ Session memories queryable immediately (no rebuild needed)
  • ✅ Deduped memories by content hash (idempotent)
  • ✅ Works with both FAISS (sidecar JSON) and Qdrant (filter-based deletion)

Key Changes from v0.4.0:

  • ✅ Memory tool now preserves session data across force rebuilds
  • ✅ Better incremental indexing for memory-heavy workflows
  • ✅ Improved metadata tracking for memory documents

Key Changes from v0.3.0:

  • ✅ SSE/HTTP Transport for daemon mode
  • ✅ OpenCode first-class support
  • ✅ Auto config generation for all clients
  • ✅ Memory tool for session persistence

Vector Store Configuration

Prometh Cortex supports two vector store backends:

Option 1: FAISS (Default - Local Storage)

Best for: Local development, private deployments, no external dependencies

[vector_store]
type = "faiss"

[storage]
rag_index_dir = ".rag_index"

Advantages:

  • ✅ No external dependencies
  • ✅ Fast local queries
  • ✅ Works offline
  • ✅ Simple setup

Disadvantages:

  • ❌ Limited to single machine
  • ❌ No concurrent write access
  • ❌ Manual backup required

Option 2: Qdrant (Cloud-native Vector Database)

Best for: Production deployments, team collaboration, scalable solutions

Local Qdrant with Docker

# Start Qdrant container with persistent storage
docker run -d \
  --name qdrant \
  -p 6333:6333 \
  -v $(pwd)/qdrant_storage:/qdrant/storage \
  qdrant/qdrant

[vector_store]
type = "qdrant"

[vector_store.qdrant]
host = "localhost"
port = 6333
collection_name = "prometh_cortex"

Cloud Qdrant

[vector_store]
type = "qdrant"

[vector_store.qdrant]
host = "your-cluster.qdrant.io"
port = 6333
collection_name = "prometh_cortex"
api_key = "your-api-key-here"
use_https = true

Advantages:

  • ✅ Concurrent access support
  • ✅ Built-in clustering and replication
  • ✅ Advanced filtering capabilities
  • ✅ REST API access
  • ✅ Automatic backups (cloud)
  • ✅ Horizontal scaling

Disadvantages:

  • ❌ Requires external service
  • ❌ Network dependency
  • ❌ Additional complexity

Qdrant Setup Steps

  1. Local Docker Setup:

    # Create persistent storage directory
    mkdir -p qdrant_storage
    
    # Start Qdrant container
    docker run -d \
      --name qdrant \
      --restart unless-stopped \
      -p 6333:6333 \
      -p 6334:6334 \
      -v $(pwd)/qdrant_storage:/qdrant/storage \
      qdrant/qdrant
    
    # Verify Qdrant is running
    curl http://localhost:6333/health
    
  2. Configure Environment:

    </code></pre>
    </li>
    </ol>
    <h1>Add to your config.toml:</h1>
    <p>[vector_store]
    type = "qdrant"</p>
    <p>[vector_store.qdrant]
    host = "localhost"
    port = 6333
    collection_name = "prometh_cortex"</p>
    <h1>api_key = ""  # Optional for local Docker</h1>
    <h1>use_https = false  # Default for local</h1>
    <pre><code>
    3. **Build Index**:
    ```bash
    # Initial build or incremental update
    pcortex build
    
    # Force complete rebuild
    pcortex rebuild --confirm
    
    1. Verify Setup:
      # Check health and statistics
      pcortex query "test" --max-results 1
      
      # Or directly check Qdrant
      curl http://localhost:6333/collections/prometh_cortex
      

    Qdrant Cloud Setup

    1. Create Qdrant Cloud Account:

      • Visit Qdrant Cloud
      • Create a cluster and get your credentials
    2. Configure Environment:

      </code></pre>
      </li>
      </ol>
      <h1>Add to your config.toml:</h1>
      <p>[vector_store]
      type = "qdrant"</p>
      <p>[vector_store.qdrant]
      host = "your-cluster-id.qdrant.io"
      port = 6333
      collection_name = "prometh_cortex"
      api_key = "your-api-key"
      use_https = true</p>
      <pre><code>
      #### Migration Between Vector Stores
      
      ```bash
      # Backup current index (if using FAISS)
      pcortex build --backup /tmp/backup_$(date +%Y%m%d_%H%M%S)
      
      # Change vector store type in config.toml
      sed -i 's/type = "faiss"/type = "qdrant"/' config.toml
      
      # Rebuild index with new vector store
      pcortex rebuild --confirm
      
      # Verify migration successful
      pcortex query "test migration" --max-results 1
      

      CLI Commands

      Build Index

      # Initial build (automatic incremental updates)
      pcortex build
      
      # Force complete rebuild (ignores incremental changes)
      pcortex build --force
      
      # Disable incremental indexing
      pcortex build --no-incremental
      
      # Rebuild entire index (with confirmation)
      pcortex rebuild
      pcortex rebuild --confirm  # Skip confirmation prompt
      

      Query Index (Unified Collection with Optional Source Filtering)

      # Query across all sources in unified collection
      pcortex query "search term"
      
      # Query with source filtering (optional)
      pcortex query "meeting notes" --source meetings
      pcortex query "action items" -s todos
      
      # Query with options
      pcortex query "search term" --max-results 5 --show-content
      

      List Sources (v0.3.0+)

      # List all configured sources with statistics
      pcortex sources
      
      # Verbose output with chunk configuration details
      pcortex sources -v
      

      Start Servers

      MCP Server (for Claude Desktop, OpenCode, Claude Code)

      # Start MCP server with stdio protocol (default)
      pcortex mcp start
      
      # Start as persistent SSE daemon (v0.4.0+)
      pcortex mcp start --transport sse --port 3100
      
      # SSE on all interfaces (for Tailscale/remote access)
      pcortex mcp start -t sse --host 0.0.0.0 -p 3100
      
      # Streamable HTTP transport (newer MCP spec)
      pcortex mcp start -t streamable-http --port 3100
      

      Generate Client Configs (v0.4.0+)

      # Generate config for various clients
      pcortex mcp init claude                        # Claude Desktop (stdio)
      pcortex mcp init opencode                      # OpenCode (stdio)
      pcortex mcp init opencode --write              # Write directly to config file
      
      # Generate SSE client configs (for daemon mode)
      pcortex mcp init claude -t sse                 # Claude Desktop (SSE)
      pcortex mcp init opencode -t sse               # OpenCode (SSE)
      pcortex mcp init opencode -t sse --url http://mac-mini.tail:3100  # Remote SSE
      

      HTTP Server (for web integrations)

      # Start HTTP server (default: localhost:8080)
      pcortex serve
      
      # Custom host/port
      pcortex serve --host 0.0.0.0 --port 9000
      
      # Development mode with auto-reload
      pcortex serve --reload
      

      Server Types (v0.3.0+, Transports v0.4.0+, Memory v0.5.0+)

      MCP Protocol Server (pcortex mcp start)

      For Claude Desktop, OpenCode, Claude Code, and other MCP clients

      Provides MCP tools with configurable transport (v0.4.0+):

      • stdio (default): Subprocess per client session, suitable for Claude Desktop
      • sse: Persistent daemon with Server-Sent Events, shared across multiple clients
      • streamable-http: Newer MCP spec HTTP transport (v0.5.0+)

      MCP Tools:

      • prometh_cortex_query: Search unified index with optional source_type filtering
      • prometh_cortex_list_sources: List all sources with statistics (v0.3.0+)
      • prometh_cortex_health: Get system health status and unified collection metrics
      • prometh_cortex_memory: Store session summaries, decisions, patterns directly to index (v0.5.0+)

      Memory Tool (v0.5.0+)

      Store and query session insights without rebuilding the entire index.

      Purpose: Capture high-value knowledge from agent sessions (OpenCode, Claude Desktop) and make it immediately searchable across your knowledge base.

      Key Features:

      • Immediate Availability: Documents queryable right after creation (no rebuild needed)
      • Automatic Deduplication: Same title + content = same document ID (idempotent)
      • Memory Preservation: Memories survive pcortex build --force and pcortex rebuild
      • Metadata Rich: Store tags, session IDs, project references, custom metadata
      • Virtual Source: Auto-injected prmth_memory source (no file-based routing)

      Parameters:

      {
        "title": "string (required) — Document title for search",
        "content": "string (required) — Markdown body (Content, Decisions, Patterns, etc.)",
        "tags": ["array of strings (optional) — e.g., 'kubernetes', 'incident', 'session'"],
        "metadata": {
          "source_project": "string (optional) — Project or context",
          "author": "string (optional) — Author/agent name",
          "session_id": "string (optional) — Session identifier",
          "custom_field": "any (optional) — Custom metadata"
        }
      }
      

      Usage Example (Claude Desktop / OpenCode):

      User: "Save this session summary to memory"
      
      Agent Response:
      prometh_cortex_memory(
        title="Session: Microservices Architecture Review - 2026-04-20",
        content="""## Summary
      Reviewed and documented the microservices architecture decisions for the platform migration project.
      
      ## Decisions Made
      - Use event-driven architecture for service communication
      - Implement circuit breaker pattern for resilience
      - Store session state in distributed cache (Redis/Memcached)
      
      ## Lessons Learned
      - Service mesh complexity grows with cluster size
      - Proper monitoring critical before production deployment
      - Version compatibility matrix must be maintained
      
      ## Next Steps
      - Document API contracts for all services
      - Set up distributed tracing infrastructure
      - Schedule follow-up architecture review in 2 weeks
      """,
        tags=["session", "architecture", "microservices"],
        metadata={
          "session_id": "sess_arch_review_2026_04_20",
          "project": "platform-migration",
          "version": "v0.5.0"
        }
      )
      

      Query Memory Documents:

      # Query across memory documents only
      pcortex query "circuit breaker pattern" --source prmth_memory
      
      # Query everywhere (memories + other sources)
      pcortex query "architecture decisions"
      
      # Via HTTP API
      curl -X POST http://localhost:8001/prometh_cortex_query \
        -H "Authorization: Bearer your-token" \
        -H "Content-Type: application/json" \
        -d '{
          "query": "session decisions and lessons learned",
          "source_type": "prmth_memory",
          "max_results": 5
        }'
      

      Force Rebuild with Memory Preservation:

      # Both commands now preserve memories (v0.5.0+)
      pcortex build --force
      pcortex rebuild --confirm
      
      # Verify memories still accessible after rebuild
      pcortex query "architecture decisions" --source prmth_memory
      

      Memory Workflow (Typical Session):

      1. During Session: Capture decisions/patterns via prometh_cortex_memory() tool
      2. Immediately Queryable: Ask "What did we decide about X?" → searches memory
      3. Force Rebuild: Run pcortex build --force when source docs change
      4. Memories Preserved: Session insights survive the rebuild
      5. Long-term KB: Export memories to permanent documents when needed

      HTTP REST Server (pcortex serve)

      For Perplexity, VSCode, web integrations

      Query Endpoint

      POST /prometh_cortex_query

      {
        "query": "search term or question",
        "max_results": 10,
        "source_type": "meetings",  // Optional: filter by source (v0.3.0+)
        "filters": {
          "datalake": "notes",
          "tags": ["work", "project"]
        }
      }
      

      List Sources Endpoint (v0.3.0+)

      GET /prometh_cortex_sources

      Returns all configured sources with:

      • Source names and chunking parameters
      • Source patterns for document routing
      • Document count per source
      • Total documents in unified index
      {
        "collection_name": "prometh_cortex",
        "sources": [
          {
            "name": "knowledge_base",
            "chunk_size": 768,
            "chunk_overlap": 76,
            "source_patterns": ["docs/specs", "docs/prds"],
            "document_count": 145
          },
          {
            "name": "meetings",
            "chunk_size": 512,
            "chunk_overlap": 51,
            "source_patterns": ["meetings"],
            "document_count": 89
          }
        ],
        "total_sources": 2,
        "total_documents": 412
      }
      

      Health Endpoint

      GET /prometh_cortex_health

      Returns server status, unified collection metrics, and performance metrics.

      Supported YAML Frontmatter Schema

      ---
      title: Document Title
      created: YYYY-MM-DDTHH:MM:SS
      author: Author Name
      category: #Category
      tags:
        - #tag1
        - tag2
      focus: Work
      uuid: document-uuid
      project:
        - name: Project Name
          uuid: project-uuid            # UUID preserved for document linking
      reminder:
        - subject: Reminder Text
          uuid: reminder-uuid           # UUID preserved for document linking
          list: List Name
      event:
        subject: Event Subject
        uuid: event-uuid                # UUID preserved for document linking
        shortUUID: MF042576B            # Short UUID also preserved
        organizer: Organizer Name
        attendees:
          - Attendee 1
          - Attendee 2
        location: Event Location
        start: YYYY-MM-DDTHH:MM:SS      # Event start time
        end: YYYY-MM-DDTHH:MM:SS        # Event end time
      related:
        - Related Item 1
        - Related Item 2
      ---
      

      Note on UUIDs for Document Linking:

      • Project, reminder, and event UUIDs are preserved in vector store metadata
      • These UUIDs enable cross-document linking and relationship queries
      • Use these UUIDs to find related documents across your datalake
      • Query by UUID: event_uuid:B897515C-1BE9-41B6-8423-3988BE0C9E3E

      YAML Frontmatter Best Practices

      ⚠️ Important: When using special characters in YAML values, always quote them properly to ensure correct parsing:

      ✅ Correct Usage:

      ---
      title: "[PRJ-0119] Add New Feature"    # Quoted because of brackets
      author: "John O'Connor"                # Quoted because of apostrophe
      tags:
        - "C#"                               # Quoted because of hash symbol
        - "project-2024"                     # Safe without quotes
      category: "Work & Personal"            # Quoted because of ampersand
      ---
      

      ❌ Problematic Usage:

      ---
      title: [PRJ-0119] Add New Feature      # Brackets will cause parsing errors
      author: John O'Connor                  # Apostrophe may cause issues
      tags:
        - C#                                 # Hash symbol conflicts with YAML
      category: Work & Personal              # Ampersand may cause issues  
      ---
      

      Common Characters That Need Quoting:

      • Square brackets []: title: "[PROJECT-123] Task Name"
      • Curly braces {}: status: "{COMPLETED}"
      • Hash/Pound #: tag: "C#"
      • Colon :: note: "Time: 3:30 PM"
      • Ampersand &: title: "Sales & Marketing"
      • Asterisk *: priority: "*HIGH*"
      • Pipe |: command: "grep | sort"
      • Greater/Less than <>: comparison: "<100ms"
      • At symbol @: email: "@company.com"
      • Apostrophes ': name: "O'Connor"

      Why This Matters:

      • Metadata Parsing: Improper YAML syntax prevents frontmatter from being extracted
      • Index Quality: Missing metadata means poor search results and filtering
      • Qdrant Storage: Malformed YAML leads to incomplete document payloads
      • Search Performance: Documents without proper metadata are harder to find

      Validation:

      Test your YAML frontmatter before indexing:

      # Quick validation of a document
      python -c "
      import yaml
      import frontmatter
      
      with open('your-document.md', 'r') as f:
          post = frontmatter.load(f)
          print('✅ YAML parsed successfully')
          print(f'Title: {post.metadata.get(\"title\", \"N/A\")}')
          print(f'Fields: {list(post.metadata.keys())}')
      "
      

      Integration

      Claude Desktop Integration

      Configure Claude Desktop by editing ~/Library/Application Support/Claude/claude_desktop_config.json:

      {
        "mcpServers": {
          "prometh-cortex": {
            "command": "/path/to/your/project/.venv/bin/python",
            "args": [
              "-m", "prometh_cortex.cli.main", "mcp"
            ],
            "env": {
              "DATALAKE_REPOS": "/path/to/your/notes,/path/to/your/documents,/path/to/your/projects",
              "RAG_INDEX_DIR": "/path/to/index/storage",
              "MCP_PORT": "8080",
              "MCP_HOST": "localhost",
              "MCP_AUTH_TOKEN": "your-secure-token",
              "EMBEDDING_MODEL": "sentence-transformers/all-MiniLM-L6-v2",
              "MAX_QUERY_RESULTS": "10",
              "CHUNK_SIZE": "512",
              "CHUNK_OVERLAP": "50",
              "VECTOR_STORE_TYPE": "faiss"
            }
          }
        }
      }
      

      Setup Steps:

      1. Install in Virtual Environment:

        cd /path/to/prometh-cortex
        python -m venv .venv
        source .venv/bin/activate  # On macOS/Linux
        pip install -e .
        
      2. Configure Settings: Create and customize your configuration:

        # Create configuration file
        cp config.toml.sample config.toml
        
        # Edit config.toml with your specific paths and settings
        # Update the [datalake] repos array with your document directories
        # Set your preferred [storage] rag_index_dir location
        # Customize [server] auth_token for security
        
      3. Build your index:

        source .venv/bin/activate
        pcortex build --force
        
      4. Get Absolute Paths: Update the MCP configuration with your actual paths:

        # Get your virtual environment Python path
        which python  # While .venv is activated
        
        # Get your project directory
        pwd
        
      5. Update Claude Desktop Config: Use absolute paths in your claude_desktop_config.json:

        {
          "mcpServers": {
            "prometh-cortex": {
              "command": "/path/to/your/project/.venv/bin/python",
              "args": [
                "-m", "prometh_cortex.cli.main", "mcp"
              ],
              "env": {
                "DATALAKE_REPOS": "/path/to/your/notes,/path/to/your/documents,/path/to/your/projects",
                "RAG_INDEX_DIR": "/path/to/index/storage",
                "MCP_PORT": "8080",
                "MCP_HOST": "localhost",
                "MCP_AUTH_TOKEN": "your-secure-token",
                "EMBEDDING_MODEL": "sentence-transformers/all-MiniLM-L6-v2",
                "MAX_QUERY_RESULTS": "10",
                "CHUNK_SIZE": "512",
                "CHUNK_OVERLAP": "50"
              }
            }
          }
        }
        
      6. Verify Configuration:

        # Test MCP server manually
        source .venv/bin/activate
        pcortex mcp  # Should start without errors
        
      7. Restart Claude Desktop: Completely quit and restart Claude Desktop application.

      Troubleshooting:

      • Check Logs: Look at Claude Desktop console logs for MCP connection errors
      • Verify Paths: Ensure all paths in the config are absolute and correct
      • Test Index: Run pcortex query "test" to verify your index works
      • Environment: Make sure environment variables are accessible from the MCP context

      Usage: After restarting Claude Desktop, you'll have access to these MCP tools:

      • prometh_cortex_query: Search your indexed documents
        • Ask: "Search my notes for yesterday's meetings"
        • Ask: "Find documents about project planning"
        • Ask: "What meetings did I have last week?"
      • prometh_cortex_health: Check system status
        • Ask: "How many documents are indexed in prometh-cortex?"
        • Ask: "What's the health status of my knowledge base?"

      OpenCode Integration

      Generate and install configuration automatically:

      # Generate OpenCode config (prints to console)
      pcortex mcp init opencode
      
      # Write directly to ~/.config/opencode/opencode.json
      pcortex mcp init opencode --write
      
      # SSE mode (requires running daemon, see SSE Daemon Mode below)
      pcortex mcp init opencode --transport sse
      

      Manual Configuration: Add the "mcp" section to ~/.config/opencode/opencode.json:

      {
        "mcp": {
          "prometh-cortex": {
            "type": "local",
            "command": ["/path/to/pcortex", "mcp", "start"],
            "environment": {
              "RAG_INDEX_DIR": "/path/to/index/storage",
              "VECTOR_STORE_TYPE": "qdrant",
              "QDRANT_HOST": "your-cluster.qdrant.io",
              "QDRANT_PORT": "6333",
              "QDRANT_COLLECTION_NAME": "prometh_cortex",
              "QDRANT_API_KEY": "your-api-key",
              "QDRANT_USE_HTTPS": "true",
              "MCP_AUTH_TOKEN": "your-secure-token",
              "EMBEDDING_MODEL": "sentence-transformers/all-MiniLM-L6-v2",
              "MAX_QUERY_RESULTS": "10"
            },
            "enabled": true,
            "timeout": 60000
          }
        }
      }
      

      Remote Mode (for persistent SSE daemon):

      {
        "mcp": {
          "prometh-cortex": {
            "type": "remote",
            "url": "http://127.0.0.1:3100/sse",
            "headers": {
              "Authorization": "Bearer your-secure-token"
            },
            "enabled": true,
            "timeout": 60000
          }
        }
      }
      

      SSE Daemon Mode

      Run Cortex as a persistent daemon instead of spawning per client session (v0.4.0+). This gives you single startup cost, shared Qdrant connections, and no duplicate vector index loads.

      1. Start the daemon:

      pcortex mcp start --transport sse --port 3100
      
      # Or bind to all interfaces for Tailscale/remote access
      pcortex mcp start -t sse --host 0.0.0.0 -p 3100
      

      2. Configure clients to connect via SSE:

      # Claude Code
      claude mcp add --transport sse prometh-cortex http://127.0.0.1:3100/sse
      
      # OpenCode
      pcortex mcp init opencode -t sse --write
      
      # Claude Desktop
      pcortex mcp init claude -t sse --write
      
      # Remote access (e.g., via Tailscale)
      pcortex mcp init opencode -t sse --url http://mac-mini.tail:3100
      

      3. (Optional) Run as macOS launchd service:

      Create ~/Library/LaunchAgents/sh.prometh.cortex-mcp.plist for auto-start on boot with keepalive.

      Claude.ai Web Integration

      Configure Claude.ai to use your MCP server by adding it as a custom integration:

      1. Start your MCP server: pcortex serve
      2. Use the webhook URL: http://localhost:8080/prometh_cortex_query
      3. Set authentication header: Authorization: Bearer your-secret-token
      4. Send queries in JSON format:
        {
          "query": "search term",
          "max_results": 10
        }
        

      Perplexity Integration

      Configure Perplexity to use your local MCP server for document search:

      Prerequisites:

      1. Start HTTP Server (not MCP protocol):

        source .venv/bin/activate
        pcortex serve --port 8001  # Use different port than MCP
        
      2. Configure for Performance (important for Perplexity timeouts):

        </code></pre>
        </li>
        </ol>
        <h1>Edit your config.toml for faster responses</h1>
        <h1>In the [embedding] section, set:</h1>
        <h1>max_query_results = 3  # Reduce from default 10 to 3</h1>
        <pre><code>
        3. **Verify Health**:
        ```bash
        curl -H "Authorization: Bearer your-secret-token" \
             http://localhost:8001/prometh_cortex_health
        

        Integration Setup:

        1. Server Configuration:

          • Protocol: HTTP
          • URL: http://localhost:8001/prometh_cortex_query
          • Method: POST
          • Headers: Authorization: Bearer your-secret-token
          • Content-Type: application/json
        2. Query Format:

          {
            "query": "your search query",
            "max_results": 3
          }
          
        3. Example Request:

          curl -X POST http://localhost:8001/prometh_cortex_query \
            -H "Authorization: Bearer your-secret-token" \
            -H "Content-Type: application/json" \
            -d '{"query": "meeting notes", "max_results": 3}'
          

        Performance Optimization:

        • Reduced Results: Use max_results: 3 instead of 10 to avoid timeouts
        • Dedicated Port: Use separate port (8001) for Perplexity vs other integrations
        • Quick Queries: Response time optimized to <400ms for timeout compatibility

        Usage in Perplexity:

        • Ask: "Search my local documents for project updates"
        • Ask: "Find my notes about last week's meetings"
        • Ask: "What information do I have about [specific topic]?"

        VSCode with GitHub Copilot Integration

        Configure VSCode to use your MCP server with GitHub Copilot:

        Option 1: VSCode MCP Extension (Recommended)

        1. Install MCP for VSCode:

          # Install the VSCode MCP extension
          code --install-extension ms-vscode.mcp
          
        2. Configure MCP Settings: Add to your VSCode settings.json or create .vscode/mcp.json:

          {
            "mcpServers": {
              "prometh-cortex": {
                "command": "/path/to/your/project/.venv/bin/python",
                "args": [
                  "-m", "prometh_cortex.cli.main", "mcp"
                ],
                "env": {
                  "DATALAKE_REPOS": "/path/to/your/notes,/path/to/your/documents,/path/to/your/projects",
                  "RAG_INDEX_DIR": "/path/to/index/storage",
                  "MCP_PORT": "8080",
                  "MCP_HOST": "localhost",
                  "MCP_AUTH_TOKEN": "your-secure-token",
                  "EMBEDDING_MODEL": "sentence-transformers/all-MiniLM-L6-v2",
                  "MAX_QUERY_RESULTS": "10",
                  "CHUNK_SIZE": "512",
                  "CHUNK_OVERLAP": "50"
                }
              }
            }
          }
          
        3. Update User Settings: Add to your VSCode settings.json:

          {
            "mcp.servers": {
              "prometh-cortex": {
                "enabled": true
              }
            }
          }
          
        4. Verify Integration:

          • Open Command Palette (Cmd+Shift+P)
          • Run "MCP: List Servers"
          • You should see "prometh-cortex" listed and active

        Option 2: Direct HTTP Integration

        Add to your VSCode settings.json:

        {
          "github.copilot.advanced": {
            "debug.useElectronPrompts": true,
            "debug.useNodeUserForPrompts": true
          },
          "prometh-cortex.server.url": "http://localhost:8001",
          "prometh-cortex.server.token": "your-secret-token"
        }
        

        Start the HTTP server:

        source .venv/bin/activate
        pcortex serve --port 8001
        

        Option 3: Custom Task Integration

        Create .vscode/tasks.json for quick queries:

        {
          "version": "2.0.0",
          "tasks": [
            {
              "label": "Query Prometh-Cortex",
              "type": "shell",
              "command": "curl",
              "args": [
                "-H", "Authorization: Bearer your-secret-token",
                "-H", "Content-Type: application/json",
                "-d", "{\"query\": \"${input:searchQuery}\", \"max_results\": 5}",
                "http://localhost:8001/prometh_cortex_query"
              ],
              "group": "build",
              "presentation": {
                "echo": true,
                "reveal": "always",
                "panel": "new"
              }
            }
          ],
          "inputs": [
            {
              "id": "searchQuery",
              "description": "Enter your search query",
              "default": "meeting notes",
              "type": "promptString"
            }
          ]
        }
        

        Setup Steps:

        1. Build Index: Ensure your RAG index is built and up-to-date

          source .venv/bin/activate
          pcortex build --force
          
        2. Start MCP Server (for Option 1):

          # MCP server runs automatically when VSCode starts
          # Check VSCode Output panel for MCP logs
          
        3. Start HTTP Server (for Options 2-3):

          source .venv/bin/activate
          pcortex serve --port 8001
          

        Usage:

        • Option 1: Use MCP commands directly in GitHub Copilot chat
          • Ask: "Search my documents for project planning notes"
          • Ask: "Find my meeting notes from last week"
        • Option 2: GitHub Copilot will automatically query your local documents
        • Option 3: Press Ctrl+Shift+P → "Tasks: Run Task" → "Query Prometh-Cortex"

        Troubleshooting:

        • Check MCP Output: View "Output" panel in VSCode, select "MCP" from dropdown
        • Verify Paths: Ensure all paths are absolute and accessible
        • Test Manually: Run pcortex mcp or pcortex serve to verify functionality
        • Restart VSCode: After configuration changes, restart VSCode completely

        Usage: Press Ctrl+Shift+P → "Tasks: Run Task" → "Query Prometh-Cortex"

        General MCP Configuration Guide

        Two Server Types Available:

        1. MCP Protocol Server (pcortex mcp start):

          • Purpose: AI assistant integration (Claude Desktop, OpenCode, Claude Code, VSCode)
          • Transports (v0.4.0+):
            • stdio (default): Subprocess per client, no network port
            • sse: Persistent daemon on configurable host:port, shared across clients
            • streamable-http: Newer MCP spec transport
          • Usage: Direct integration with MCP-compatible clients
        2. HTTP REST Server (pcortex serve):

          • Purpose: Web applications, HTTP clients (Perplexity, custom integrations)
          • Protocol: HTTP REST API
          • Port: Configurable (default: 8080)
          • Usage: Traditional HTTP API access

        MCP Transport Selection Guide

        Transport Best For Port Startup Shared State Setup
        stdio Single client (Claude Desktop) None ~2s No Simple: pcortex mcp
        sse Multiple clients (OpenCode + Claude Desktop) Yes ~2s Yes Daemon: pcortex mcp start -t sse
        streamable-http HTTP clients + MCP Yes ~2s Yes Daemon: pcortex mcp start -t streamable-http

        Decision Tree:

        • Just using Claude Desktop? → Use stdio (default)
        • Using OpenCode + Claude Desktop? → Use sse daemon (shared startup cost)
        • Remote access needed? → Use sse with --host 0.0.0.0 (Tailscale/SSH tunnel)
        • Need HTTP API + MCP? → Use streamable-http daemon

        Configuration Prerequisites:

        1. Environment Setup:

          # Create and activate virtual environment
          python -m venv .venv
          source .venv/bin/activate  # macOS/Linux
          
          # Install in development mode
          pip install -e .
          
        2. Create Configuration:

          # Create configuration from sample
          cp config.toml.sample config.toml
          # Edit config.toml with your specific settings
          
        3. Build Index:

          pcortex build --force
          
        4. Test Configuration:

          # Test MCP server
          pcortex mcp  # Should start without errors, Ctrl+C to stop
          
          # Test HTTP server
          pcortex serve  # Should show server info, Ctrl+C to stop
          
          # Test query functionality
          pcortex query "test search"
          

        Common Integration Pattern:

        For HTTP integrations (Perplexity, web apps):

        # Start HTTP server
        pcortex serve --port 8001
        
        # Query endpoint
        POST http://localhost:8001/prometh_cortex_query
        Authorization: Bearer your-secret-token
        Content-Type: application/json
        
        {
          "query": "your search query",
          "max_results": 10,
          "filters": {
            "datalake": "notes",
            "tags": ["work"]
          }
        }
        
        # Health check
        GET http://localhost:8001/prometh_cortex_health
        Authorization: Bearer your-secret-token
        

        For MCP integrations (Claude Desktop, VSCode):

        {
          "mcpServers": {
            "prometh-cortex": {
              "command": "/path/to/your/project/.venv/bin/python",
              "args": [
                "-m", "prometh_cortex.cli.main", "mcp"
              ],
              "env": {
                "DATALAKE_REPOS": "/path/to/your/notes,/path/to/your/documents,/path/to/your/projects",
                "RAG_INDEX_DIR": "/path/to/index/storage",
                "MCP_PORT": "8080",
                "MCP_HOST": "localhost",
                "MCP_AUTH_TOKEN": "your-secure-token",
                "EMBEDDING_MODEL": "sentence-transformers/all-MiniLM-L6-v2",
                "MAX_QUERY_RESULTS": "10",
                "CHUNK_SIZE": "512",
                "CHUNK_OVERLAP": "50",
                "VECTOR_STORE_TYPE": "faiss"
              }
            }
          }
        }
        

        Performance Tuning:

        • For Perplexity: Set max_query_results = 3 in config.toml to avoid timeouts
        • For Development: Use --reload flag with pcortex serve
        • For Production: Use production WSGI server instead of development server

        Auto-start Script: Create start_servers.sh for easy management:

        #!/bin/bash
        # Kill existing servers
        pkill -f "pcortex serve" 2>/dev/null || true
        pkill -f "pcortex mcp" 2>/dev/null || true
        
        # Activate virtual environment
        source .venv/bin/activate
        
        # Start HTTP server in background
        nohup pcortex serve --port 8001 > /tmp/prometh-cortex-http.log 2>&1 &
        
        echo "Prometh-Cortex servers started"
        echo "HTTP Server: http://localhost:8001"
        echo "MCP Server: Available for stdio connections"
        echo "Logs: /tmp/prometh-cortex-http.log"
        

        Troubleshooting Checklist:

        • Virtual Environment: Always use absolute paths to .venv/bin/python
        • Configuration: Set datalake.repos and storage.rag_index_dir in config.toml
        • Index Built: Run pcortex build before using servers
        • Ports Available: Check port conflicts with lsof -i :8080
        • Logs Check: Monitor server logs for configuration errors
        • Path Permissions: Ensure read access to datalake and write access to index directory

        Development

        Setup Development Environment

        # Clone repository
        git clone https://github.com/prometh-sh/prometh-cortex.git
        cd prometh-cortex
        
        # Install with development dependencies
        pip install -e ".[dev]"
        
        # Install pre-commit hooks
        pre-commit install
        

        Run Tests

        # Run all tests
        pytest
        
        # Run with coverage
        pytest --cov=src/prometh_cortex
        
        # Run specific test types
        pytest tests/unit/
        pytest tests/integration/
        

        Code Quality

        # Format code
        black src/ tests/
        isort src/ tests/
        
        # Lint code
        flake8 src/ tests/
        
        # Type checking
        mypy src/
        

        Performance

        • Query Speed: Target <100ms on M1/M2 Mac
        • Index Size: Scales to thousands of documents
        • Memory Usage: Optimized chunking and streaming processing
        • Storage: Efficient FAISS local storage or scalable Qdrant
        • Incremental Updates: Only processes changed documents

        Architecture

        ┌─────────────────────┐
        │    config.toml      │
        └──────────┬──────────┘
                   │
        ┌──────────▼──────────────────┐
        │ Datalake Ingest & Parser    │
        │ - Markdown files            │
        │ - YAML frontmatter          │
        └──────────┬──────────────────┘
                   │
        ┌──────────▼──────────────────┐
        │ Vector Store / Indexing     │
        │ - FAISS (local) or Qdrant   │
        │ - Local embedding model     │
        │ - Incremental indexing      │
        └──────────┬──────────────────┘
                   │
        ┌──────────▼──────────────────┐
        │     MCP Server              │
        │ - stdio / SSE / HTTP        │
        │ - prometh_cortex_query      │
        │ - prometh_cortex_health     │
        │ - prometh_cortex_sources    │
        └──────────┬──────────────────┘
                   │
            ┌──────┼──────┐
            │      │      │
          stdio   SSE   HTTP
            │      │      │
         Claude  Multi   REST
        Desktop client  API
                 daemon
        

        License

        Apache 2.0 License - see LICENSE for details.

        Contributing

        We welcome contributions! Please see CONTRIBUTING.md for detailed guidelines.

        Quick Contribution Guide

        1. Fork the repository
        2. Create a feature branch: git checkout -b feature/your-feature-name
        3. Make your changes with clear, descriptive commits
        4. Add tests for new functionality
        5. Ensure all tests pass: pytest
        6. Format code: black src/ tests/ and isort src/ tests/
        7. Submit a pull request with a clear description

        Code of Conduct

        This project follows our Code of Conduct. By participating, you agree to uphold this code.

        Security

        Found a security vulnerability? Please see SECURITY.md for responsible disclosure guidelines.

        Documentation

        Architecture & Design:

        Migration Guides:

        Key Improvements in v0.5.0:

        • Memory Preservation: Session memories survive pcortex build --force and pcortex rebuild
        • Memory Tool (MCP): prometh_cortex_memory() for capturing decisions, patterns, and session summaries
        • Dual Backend Support: FAISS (sidecar JSON) and Qdrant (filter-based) memory preservation
        • Smart Metadata Retrieval: Handle both parent document IDs and chunk IDs seamlessly

        Key Improvements in v0.4.0:

        • SSE/HTTP Transport: Run MCP as a persistent daemon shared across clients
        • OpenCode Support: First-class config generation for OpenCode
        • Auto Config: pcortex mcp init <target> generates configs for Claude, OpenCode, VSCode, Codex, Perplexity
        • Remote Access: SSE daemon with --host 0.0.0.0 for Tailscale/multi-machine setups

        Key Improvements in v0.3.0:

        • Unified Collection: Single FAISS/Qdrant index instead of multiple
        • Per-Source Chunking: Different chunk sizes per document source in unified index
        • Topic-Based Queries: Query across document types naturally
        • Better Performance: ~300ms queries (vs ~500ms multi-collection)
        • Lower Memory: Single index (vs 3-5x for multi-collection)

        Support

        Getting Help

        Resources

        Community

        We encourage community participation! Whether you're fixing bugs, adding features, improving documentation, or helping others, all contributions are valued.


        Made with ❤️ for the knowledge management community

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

prometh_cortex-0.5.1.tar.gz (172.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

prometh_cortex-0.5.1-py3-none-any.whl (135.8 kB view details)

Uploaded Python 3

File details

Details for the file prometh_cortex-0.5.1.tar.gz.

File metadata

  • Download URL: prometh_cortex-0.5.1.tar.gz
  • Upload date:
  • Size: 172.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.15

File hashes

Hashes for prometh_cortex-0.5.1.tar.gz
Algorithm Hash digest
SHA256 ffbacdbe1d3d37091458916befb92f664f79b0589bd920c35f21c752253fde0c
MD5 5c5d999f10d4c79c3d13a3655b99d387
BLAKE2b-256 0582cce48aab4a2e273ee44d94d61ef1eb0323f6109898654b91b4fc995bfe9d

See more details on using hashes here.

File details

Details for the file prometh_cortex-0.5.1-py3-none-any.whl.

File metadata

  • Download URL: prometh_cortex-0.5.1-py3-none-any.whl
  • Upload date:
  • Size: 135.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.15

File hashes

Hashes for prometh_cortex-0.5.1-py3-none-any.whl
Algorithm Hash digest
SHA256 3fd6e2b5f6b2eb43d2787fb67c1be5824ff50c2e1dbcc4f480083be17b2518a7
MD5 a0833905be89fc50470c8f664f49da5f
BLAKE2b-256 0b6dee04ff63ea2c20138f2a14d455398f8714f0946c84b1100aec63b16c200b

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page