Skip to main content

Temporal Layered Context Memory Engine - Biological Edge-Optimized memory for AI Agents

Project description

Temporal Layered Context Memory (TLCM) Engine

"Current AI memory systems treat memory as a filing cabinet: store, retrieve, overwrite. The human brain treats memory as a living architecture: versioned, temporally indexed, emotionally weighted, and spatially separated by context."Collins Somtochukwu (Harper Kollins)

TLCM is a neuro-inspired AI memory engine built to solve the structural failures of current Large Language Model memory architectures. Instead of relying on sprawling, chaotic vector dumps or flat un-versioned updates, TLCM implements time and context as fundamental, first-class architectural dimensions.

Tests Version License


The Structural Problem

The TLCM Engine solves three exact failure modes present in modern LLMs:

  1. The Snapshot Problem: AI systems store facts as static photographs. When reality updates, the photograph becomes fiction.
  2. The Overwrite Problem: When current AI does update, it destroys the old memory. It loses the evolutionary arc. It cannot answer: "How did we get from what we believed in 2024 to what we believe in 2026?"
  3. The Context Bleed Problem: AIs process all active goals in one flat namespace. Your startup data blends with your personal relationship data, creating unauthorized, hallucinatory connections.

How TLCM Solves It — The 4 Principles

TLCM maps cognitive neuroscience directly to software engineering:

Principle 1: Versioned Memory (No Overwrite)

Powered by specialized SQLite version-chain logic, memories are never deleted. When a fact updates, a new version is created containing a pointer (parent_id) to the previous version and a reason for the shift. The system operates as a Git repository for AI beliefs.

Principle 2: Temporal Epoch Tagging

Inspired by autobiographical "lifetime periods," every fact belongs to a contextual epoch.

Principle 3: Context Workspace Isolation

Powered by strictly segregated ChromaDB semantic namespaces, cognitive workspaces are isolated. A query about your screenplay is mathematically incapable of retrieving vectors regarding your AI startup unless you authorize an explicit cross-workspace link.

Principle 4: The Temporal Jump (Mathematical Semantic Delta)

When asked to evaluate the past, TLCM reconstructs a complete world-state using the preserved memory graphs. The algorithm calculates the delta in pure Python.


Neuro-Weighted Biological Decay Algorithm (v0.5)

Inspired by the Ebbinghaus Forgetting Curve and extended with emotional reconsolidation theory, TLCM implements a neuro-weighted decay mechanism where emotional intensity and urgency directly affect how fast memories fade.

The Formal Decay Equation

\mathcal{C}_{t} = \max\left( 0.1, \mathcal{C}_{t-1} - \frac{\alpha}{1 + \frac{\mathcal{U}}{10} + \frac{|\mathcal{E}|}{10}} \right)

Where:

  • $\mathcal{C}_{t}$ is Confidence at time $t$
  • $\alpha$ is the base decay rate (e.g., 0.05)
  • $\mathcal{U}$ is Urgency Score $(0 - 10)$
  • $\mathcal{E}$ is Emotional Valence $(-10 \dots +10)$

The decay daemon runs automatically as a periodic background task (default: 24h cycle, configurable via TLCM_DECAY_INTERVAL_SECONDS). Manual trigger available at POST /api/v1/decay/run.


Surprise-Driven Reconsolidation

When a new high-urgency memory ($\mathcal{U} > 7$) arrives, the background daemon automatically surfaces semantically related older memories and actively boosts their confidence ($\mathcal{C}{t} = \mathcal{C}{t} + 0.1$). This is pure neuroscience replication: highly charged new events reinforce associated older memories.


Pluggable Cognition Architecture

TLCM's cognition backend can be seamlessly toggled between cloud and air-gapped operation:

Backend Provider Use Case
gemini (default) Gemini 3.1 Flash Lite Cloud, high-precision neuro-scoring
ollama Any local Ollama model Air-gapped, privacy-preserving, edge
# Toggle via environment variable
COGNITION_BACKEND=gemini   # Cloud mode (default)
COGNITION_BACKEND=ollama   # Air-gapped mode
OLLAMA_COGNITION_MODEL=gemma2:2b  # Choose any local model

Custom providers (OpenAI, Anthropic, etc.) can be added by subclassing CognitionProvider in core/interfaces.py.


Formal Properties

Guaranteed Workspace Isolation Under Concurrent Updates

A mathematical limit via ChromaDB metadata filtering ensures P(Retrieval | Workspace_ID_A) = 0 for Workspace_ID_B. The async memory bus manages thread-safe SQLite transactions alongside vector locks ensuring no cross-contamination under heavy multi-threading.


Why TLCM Is Different (Comparison Table)

Dimension TLCM (v0.5) Mem0 Zep / Graphiti Letta / MemGPT
Versioned History Full Git-style chains Flat Overwrite Graph Edges Mutable Archival
Workspace Isolation Mathematical Zero-Bleed Single Global Graph Tenant Windows Single Namespace
Contradiction Physics Cascade Orphaning Overwrite/Delete Edge Reweighting LLM Selection
Decay Mechanism Neuro-Weighted Math Custom/None None Eviction Policies
Semantic Delta Pre-computed Python Sets LLM Hallucinated diff LLM hallucinated diff Search
Pluggable Backend Gemini / Ollama / Custom OpenAI Only OpenAI/Azure OpenAI Only

1k LoCoMo Benchmark Results

TLCM scaled to a 1000+ deterministic node graph evaluation via the LoCoMo context benchmark framework over 200 temporal and evolutionary queries.

Metric TLCM Accuracy Observation
Point-in-Time Retrieval 98.8% Maintains exact timeline snapshots without future-state leakage.
Evolution Tracking 98.3% Successfully traverses full DAG parent chains continuously.
Contradiction Resolution 100% Cascade Orphaning flawlessly eliminates all outdated graph branches.

Unit Test Coverage (46/46 Passing)

Test Suite Tests Coverage
Memory Store 11 remember, recall, update, version chains
Graph Surgery 6 Cascade orphaning, confidence zeroing
Workspace Isolation 5 Cross-workspace zero-bleed proof
Temporal Jump 8 Delta calculation, additions detection
Decay Math 6 Ebbinghaus equation, floor, urgency/emotion resistance
Epoch Lifecycle 10 Create, list, close, active state
# Run all tests (no GPU required)
TLCM_TEST_MODE=1 python -m pytest tests/unit/ -v

Ablation Results (Feature Proof)

Configuration Retrieval Accuracy Isolation Drift/Orphan Removal Rate
TLCM Full 98.8% PASS 100% Removed
No Decay 82.5% PASS 100%
No Epochs 41.0% PASS Failed
No Math Delta 12.0% PASS Failed (Hallucinated)
No Workspace Isolation 63.5% FAIL 100%

Hardware Reality (Edge-First Design)

TLCM executes orchestration natively on edge devices leveraging an asynchronous ingestion tier MemoryBus to handle STM (Short Term Memory) spikes.

Performance on Constrained Hardware:

  • Machine: Intel i5, 16GB RAM (No GPU).
  • STM Ingestion: <20ms per query.
  • LTM Processing (Background): ~120ms asynchronous resolution.
  • Latency is successfully decoupled from LLM inference.

Quick Start Guide

1. Install Dependencies

pip install -e .

2. Configure Backend Engine

Create a .env file in the project root:

# Cognition backend (gemini or ollama)
COGNITION_BACKEND=gemini
GEMINI_API_KEY=your_key_here

# Optional: Secure the API
TLCM_API_KEY=super_secret_key

# Optional: Air-gapped mode
# COGNITION_BACKEND=ollama
# OLLAMA_COGNITION_MODEL=gemma2:2b

3. Run the Slim Node Server

python -m uvicorn server.main:app --reload --port 8000

4. Dashboard

The web dashboard is available at http://localhost:8000/dashboard (requires building the frontend first):

cd tlcm-web && npm install && npm run build

5. API Endpoints

Method Endpoint Description
POST /api/v1/memories/remember Async memory ingestion via bus
POST /api/v1/memories/remember/sync Synchronous memory storage
POST /api/v1/memories/recall Semantic memory recall
GET /api/v1/memories/{id} Fetch single memory
GET /api/v1/memories/{id}/history Version chain
POST /api/v1/workspaces/ Create workspace
DELETE /api/v1/workspaces/{name} Delete workspace
POST /api/v1/epochs/ Create epoch
POST /api/v1/jump/delta Temporal jump delta
GET /api/v1/health Health check
POST /api/v1/decay/run Manual decay trigger
GET /api/v1/events SSE event stream

SDKs

  • Python: pip install tlcm-client — see sdk/python/
  • TypeScript: npm install tlcm-client — see sdk/typescript/
  • LangChain: Native integration via integrations/langchain/
  • Letta/MemGPT: Archival memory adapter via integrations/letta/

Limitations & Future Work

We recognize several boundaries in the v0.5 architecture targeting deployment at scale:

  • Node Sync: Multi-agent temporal synchronization across distributed TLCM edge nodes requires vector clock synchronization logic.
  • Scale: Scaling beyond 2M memories per workspace may require Sharded ChromaDB clusters.
  • RL-Driven Reconsolidation: Next-generation reconsolidation will look into Reinforcement Learning directly mapped to survival heuristics.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

tlcm_engine-0.5.1.tar.gz (51.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

tlcm_engine-0.5.1-py3-none-any.whl (51.6 kB view details)

Uploaded Python 3

File details

Details for the file tlcm_engine-0.5.1.tar.gz.

File metadata

  • Download URL: tlcm_engine-0.5.1.tar.gz
  • Upload date:
  • Size: 51.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.2

File hashes

Hashes for tlcm_engine-0.5.1.tar.gz
Algorithm Hash digest
SHA256 18c9bd69edbdb374b1da7282b7e4412f2d19fc78486265639b9ae228896095e5
MD5 2429563ef2dba523cbf26e0c889c7603
BLAKE2b-256 e32d43a75488987151fdd6f0930811ce627f28accbff9ed9c08cd473148c2c75

See more details on using hashes here.

File details

Details for the file tlcm_engine-0.5.1-py3-none-any.whl.

File metadata

  • Download URL: tlcm_engine-0.5.1-py3-none-any.whl
  • Upload date:
  • Size: 51.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.2

File hashes

Hashes for tlcm_engine-0.5.1-py3-none-any.whl
Algorithm Hash digest
SHA256 354bbd68089aa6f7126e7fb7d7785ecdfc70c47dc524b0d359262f3ef6d9370e
MD5 7fe52f383995ddbb8e6867305e6f6b77
BLAKE2b-256 3e695ba9f1c0888f34f6a31c5daf135c0efab86c60e4711a80ce81c27eabeded

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page