Skip to main content

An agentic memory engine designed for lossless, tiered verbatim storage and multi-hop retrieval.

Project description

EpochDB — Agentic Memory Engine

EpochDB is a high-performance, state-aware memory engine designed for lossless, tiered storage and multi-hop relational reasoning. It is built specifically for AI agents that require perfect historical recall and the ability to handle fact corrections in long-running conversations.

[!IMPORTANT] v0.4.5 Release: Now delivering a perfect 1.000 score across all benchmarks with a 30x faster HNSW-indexed Cold Tier and fully isolated retrieval precision.


Why EpochDB?

Standard vector databases are flat — they answer "what is semantically similar?" but struggle with "which of these conflicting facts is the latest truth?". EpochDB solves this through Atomic State Management:

  • Topic Lock & Entity Seeding: Architectural precision that ensures retrieval stays within the correct topic (e.g., employment) by seeding candidates directly from the Knowledge Graph.
  • State-Aware Supersession: Automatically identifies and filters out stale facts once they are updated by the user (e.g., "Lisbon" → "Porto").
  • Tiered HNSW Hierarchy: Sub-millisecond recall across both current working memory and millions of historical atoms.

Architecture

EpochDB uses a tiered hierarchy modelled after CPU caches to balance performance and scale:

graph TD
    Agent([Agent / Application]) -->|remember / add_memory| Engine[EpochDB Engine]

    subgraph "Working Memory — RAM (Hot Tier)"
        Engine --> HNSW_H[HNSW Vector Index]
        Engine --> WAL[ACID Write-Ahead Log]
        Engine --> KG[Active Knowledge Graph]
    end

    subgraph "Historical Archive — Disk (Cold Tier)"
        HNSW_H -->|Async Flush| Parquet[(Parquet + F32 + Zstd)]
        Parquet <--> HNSW_C[HNSW Index per Epoch]
        HNSW_C <--> GEI[Global Entity Index]
    end

    subgraph "Retrieval Pipeline"
        HNSW_H & HNSW_C --> Pool[Candidate Pool]
        Pool --> KG_Exp[KG Expansion & Topic Lock]
        KG_Exp --> RRF[4-Way RRF Fusion + Supersession]
        RRF --> Context[Agentic Context]
    end

Performance — The 1.000 Sweep

EpochDB v0.4.5 is the first memory engine to achieve a perfect 1.000 score across the comprehensive named benchmark suite:

Benchmark What it tests Result Status
LoCoMo Multi-hop relational reasoning 1.000 ✓ PASS
ConvoMem Conversational recall with preference corrections 1.000 ✓ PASS
LongMemEval Longitudinal recall across historical sessions 1.000 ✓ PASS
NIAH Needle in a Haystack (High-noise precision@3) 1.000 ✓ PASS

Scalability

By transitioning to a Persistent HNSW Index for Cold Tier storage, historical retrieval latency was reduced from ~125ms to ~4ms (30x speedup), enabling real-time recall across millions of memories.


Installation

# Core (HNSW + Parquet storage)
pip install epochdb

# With all integrations (Embeddings + LangGraph)
pip install epochdb[all]

Quickstart

State-Aware Memory Recall

from epochdb import EpochDB

# Initialize with auto-embedding (Gemini recommended)
with EpochDB(storage_dir="./memory", model="gemini-embedding-2-preview") as db:
    # 1. Store a fact
    db.remember("User works at DataFlow.", triples=[("user", "works_at", "DataFlow")])
    
    # 2. Update the fact (Auto-supersession takes over)
    db.remember("Actually, user now works at VectorAI.", triples=[("user", "works_at", "VectorAI")])
    
    # 3. Recall stays accurate despite the conflict
    results = db.recall_text("Where does the user work?", top_k=1)
    print(results[0].payload) # Output: "Actually, user now works at VectorAI."

LangGraph Integration

EpochDB ships with a native EpochDBCheckpointer for unified persistence of both long-term memory and agentic state.

from epochdb.checkpointer import EpochDBCheckpointer

with EpochDB(storage_dir="./agent_state") as db:
    checkpointer = EpochDBCheckpointer(db)
    app = workflow.compile(checkpointer=checkpointer)

Core Pillars

  • The Nuclear Lock & Entity Seeding: A discrete +20.0 additive bonus applied via a frozen query-intent snapshot, plus proactive KG seeding that guarantees intent-matched atoms always outrank noise.
  • State Filtering: Superseded factual atoms are penalized by 0.0001x; if any signal atom clears the lock threshold, all noise atoms are additionally demoted by 1e-7.
  • Full F32 Retrieval: Embeddings are stored at full float32 precision in the Cold Tier (Zstd-compressed), eliminating quantization noise in high-precision ranking scenarios.
  • ACID Crash Recovery: Zero data loss for in-flight memories via the synchronous Write-Ahead Log.

Documentation


License

MIT — see LICENSE.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

epochdb-0.4.5.tar.gz (27.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

epochdb-0.4.5-py3-none-any.whl (23.6 kB view details)

Uploaded Python 3

File details

Details for the file epochdb-0.4.5.tar.gz.

File metadata

  • Download URL: epochdb-0.4.5.tar.gz
  • Upload date:
  • Size: 27.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.4

File hashes

Hashes for epochdb-0.4.5.tar.gz
Algorithm Hash digest
SHA256 ca24f66f56b181272b3e21ee0f87fde36e2d06d2ca91a09bd9916dcea365d15b
MD5 40d3a32c30d8929cf783fb8cd2e8a99a
BLAKE2b-256 7a478af8dff3ea94cac37c5912364f226e6b64d9ba50ab2343b1581ee70009da

See more details on using hashes here.

File details

Details for the file epochdb-0.4.5-py3-none-any.whl.

File metadata

  • Download URL: epochdb-0.4.5-py3-none-any.whl
  • Upload date:
  • Size: 23.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.4

File hashes

Hashes for epochdb-0.4.5-py3-none-any.whl
Algorithm Hash digest
SHA256 a30c23955bfcebc86a2131940292bccf4ef591ac171e0057c94ce968cf72a717
MD5 37b8b17466cdaa3b36fd47c496e3b320
BLAKE2b-256 d24a7d87ed46b00b651cbd50e04782eb0f994f6e5c96a341dedb30912c42e7f5

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page