File-based persistent memory for AI agents. Zero dependencies.
Project description
Antaris Memory
Production-ready file-based persistent memory for AI agents. Zero dependencies (core).
Store, search, decay, and consolidate agent memories using only the Python standard library. Sharded storage for scalability, fast search indexes, namespace isolation, MCP server support, and automatic schema migration. No vector databases, no infrastructure, no API keys.
What's New in v2.0.0
- MCP Server — expose your memory workspace as MCP resources and tools; integrate with any MCP-compatible host via
create_mcp_server()(requirespip install mcp) - Hybrid semantic search — plug in any embedding function with
set_embedding_fn(fn); BM25 and cosine similarity blend automatically when an embedding function is set - Memory types — typed memory ingestion:
episodic,semantic,procedural,preference,mistake— each with custom recall priority boosts - Namespace isolation —
NamespacedMemoryandNamespaceManagerfor multi-tenant agent memory with hard boundaries - Context packets —
ContextPacket/build_context_packet()packages relevant memories for sub-agent injection with token budgeting - LRU read cache — Sprint 11 search caching with access-count boosting; configurable size via
cache_max_entries - 293 tests (all passing)
See CHANGELOG.md for full version history.
OpenClaw Integration
antaris-memory ships as a native OpenClaw plugin (antaris-memory). Install it once and your agent automatically recalls relevant memories before each turn and ingests new context after each turn — no code changes required.
# Install the OpenClaw plugin
cp -r antaris-memory-plugin/ ~/.openclaw/extensions/antaris-memory/
# Restart OpenClaw — hooks fire automatically
What the plugin does:
before_agent_start— searches memory for relevant context, injects into agent promptagent_end— ingests the turn into persistent memory
Also ships with an MCP server for integration with any MCP-compatible host:
from antaris_memory import create_mcp_server # pip install mcp
server = create_mcp_server(workspace="./memory")
What It Does
- Sharded storage for production scalability (10,000+ memories, sub-second search)
- Fast search indexes (full-text, tags, dates) stored as transparent JSON files
- Automatic schema migration from single-file to sharded format with rollback
- Multi-agent shared memory pools with namespace isolation and access controls
- Retrieval weighted by recency × importance × access frequency (Ebbinghaus-inspired decay)
- Input gating classifies incoming content by priority (P0–P3) and drops ephemeral noise at intake
- Detects contradictions between stored memories using deterministic rule-based comparison
- Runs fully offline — zero network calls, zero tokens, zero API keys
What It Doesn't Do
- Not a vector database — no embeddings by default. Core search uses BM25 keyword ranking. Semantic search requires you to supply an embedding function (
set_embedding_fn(fn)) — we never make that call for you. - Not a knowledge graph — flat memory store with metadata indexing. No entity relationships or graph traversal.
- Not semantic by default — contradiction detection compares normalized statements using explicit conflict rules, not inference. It will not catch contradictions phrased differently.
- Not LLM-dependent — all operations are deterministic. No model calls, no prompt engineering.
- Not infinitely scalable — JSON file storage works well up to ~50,000 memories per workspace. Beyond that, you'll want a database. We're honest about this so you don't discover limits in production.
Design Goals
| Goal | Rationale |
|---|---|
| Deterministic | Same input → same output. No model variance. |
| Offline | No network, no API keys, no phoning home. |
| Minimal surface area | One class (MemorySystem), obvious method names. |
| No hidden processes | Consolidation and synthesis run only when called. |
| Transparent storage | Plain JSON files. Inspect with any text editor. |
Install
pip install antaris-memory
Quick Start
from antaris_memory import MemorySystem
mem = MemorySystem("./workspace", half_life=7.0)
mem.load() # No-op on first run
# Store memories (typed ingestion)
mem.ingest("Decided to use PostgreSQL for the database.",
source="meeting-notes", category="strategic", memory_type="episodic")
mem.ingest("API costs $500/month — too expensive.",
source="review", category="operational")
# Typed helpers
mem.ingest_fact("PostgreSQL supports JSON natively")
mem.ingest_preference("User prefers concise explanations")
mem.ingest_mistake("Forgot to close DB connections in worker threads")
mem.ingest_procedure("Deploy: push to main → CI runs → auto-deploy to staging")
# Input gating — drops ephemeral noise (P3) before storage
mem.ingest_with_gating("Decided to switch to Redis for caching", source="chat")
mem.ingest_with_gating("thanks for the update!", source="chat") # → dropped (P3)
# Search (BM25; hybrid BM25+cosine if embedding fn set)
for r in mem.search("database decision"):
print(f"[{r.confidence:.2f}] {r.content}")
# → [1.00] Decided to use PostgreSQL for the database.
# Detailed search with score explanations
for r in mem.search("database decision", explain=True):
print(f"[{r.relevance:.2f}] {r.content[:60]} ({r.explanation})")
# Temporal queries
mem.on_date("2026-02-14")
mem.narrative(topic="database migration")
# Build a context packet for sub-agent injection
packet = mem.build_context_packet(
task="Optimize database performance",
max_memories=10,
max_tokens=2000,
)
print(packet.render("markdown"))
# Selective deletion (GDPR-ready with audit trail)
mem.forget(entity="John Doe")
mem.forget(before_date="2025-01-01")
# Background consolidation
report = mem.consolidate()
mem.save()
Hybrid Semantic Search
Plug in any embedding function to blend BM25 with cosine similarity:
import openai
def my_embed(text: str) -> list[float]:
resp = openai.embeddings.create(model="text-embedding-3-small", input=text)
return resp.data[0].embedding
mem = MemorySystem("./workspace")
mem.load()
mem.set_embedding_fn(my_embed) # BM25+cosine hybrid activates automatically
# Or use a local model — anything that returns a list of floats works
import ollama
def local_embed(text: str) -> list[float]:
return ollama.embeddings(model="nomic-embed-text", prompt=text)["embedding"]
mem.set_embedding_fn(local_embed)
When no embedding function is set, search uses BM25 only (zero API calls).
Memory Types
Store memories with type-specific recall boosts:
mem.ingest("Deploy: push to main, CI runs, auto-deploy to staging",
memory_type="procedural") # High recall boost for how-to queries
mem.ingest_fact("PostgreSQL supports JSONB indexing") # Semantic memory
mem.ingest_preference("User prefers Python examples") # Preference memory
mem.ingest_mistake("Forgot to handle connection timeout") # Mistake memory
mem.ingest_procedure("Run pytest from venv, not global pip") # Procedure memory
| Type | Use for | Recall boost |
|---|---|---|
episodic |
Events, decisions, meeting notes | Normal |
semantic |
Facts, concepts, general knowledge | Medium |
procedural |
How-to steps, runbooks | High |
preference |
User preferences, style notes | High |
mistake |
Errors to avoid, lessons learned | High |
Namespace Isolation
Multi-tenant or multi-agent workspaces with hard boundaries:
from antaris_memory import NamespacedMemory, NamespaceManager
manager = NamespaceManager("./workspace")
# Each agent gets its own isolated memory
agent_a = manager.create_namespace("agent-a")
agent_b = manager.create_namespace("agent-b")
# Or use NamespacedMemory directly
ns = NamespacedMemory("project-alpha", "./workspace")
ns.load()
ns.ingest("Alpha-specific decision")
results = ns.search("decision")
Context Packets (Sub-Agent Injection)
Package relevant memories for sub-agents spawning cold:
# Single-query context packet
packet = mem.build_context_packet(
task="Debug the authentication flow",
tags=["auth", "security"],
max_memories=10,
max_tokens=2000,
include_mistakes=True,
)
print(packet.render("markdown")) # → structured markdown for prompt injection
print(packet.render("xml")) # → XML format
# Multi-query with deduplication
packet = mem.build_context_packet_multi(
task="Fix performance issues",
queries=["database bottleneck", "slow queries", "caching strategy"],
max_tokens=3000,
)
# Trim to token budget
packet.trim(max_tokens=1500)
MCP Server
Expose your memory workspace to any MCP-compatible host:
from antaris_memory import create_mcp_server # pip install mcp
server = create_mcp_server("./workspace")
server.run() # Stdio transport — connect from Claude Desktop, Cursor, etc.
MCP tools exposed: memory_search, memory_ingest, memory_consolidate, memory_stats.
Shared Memory Pools
Multiple agents sharing a common memory workspace:
from antaris_memory import SharedMemoryPool, AgentPermission
pool = SharedMemoryPool("./shared", pool_name="team-alpha")
pool.grant("agent-1", AgentPermission.READ_WRITE)
pool.grant("agent-2", AgentPermission.READ_ONLY)
# Agent 1 writes
mem_1 = pool.open("agent-1")
mem_1.ingest("Deployed new API endpoint")
# Agent 2 reads
mem_2 = pool.open("agent-2")
results = mem_2.search("API deployment")
Input Gating (P0–P3)
Classifies content at intake. Low-value data never enters storage:
mem.ingest_with_gating("CRITICAL: API key compromised", source="alerts")
# → P0 (critical) → stored
mem.ingest_with_gating("Decided to switch to PostgreSQL", source="meeting")
# → P1 (operational) → stored
mem.ingest_with_gating("thanks for the update!", source="chat")
# → P3 (ephemeral) → dropped
| Level | Category | Stored | Examples |
|---|---|---|---|
| P0 | Strategic | ✅ | Security alerts, errors, deadlines, financial commitments |
| P1 | Operational | ✅ | Decisions, assignments, technical choices |
| P2 | Tactical | ✅ | Background info, research, general discussion |
| P3 | — | ❌ | Greetings, acknowledgments, filler |
Classification: keyword and pattern matching — no LLM calls. 0.177ms avg per input.
Concurrency
Multiple processes can safely read and write to the same memory workspace.
from antaris_memory import FileLock, VersionTracker
# Exclusive write access
with FileLock("/path/to/shard.json", timeout=10.0):
data = load(shard)
modify(data)
save(shard, data)
# Optimistic conflict detection for read-heavy workloads
tracker = VersionTracker()
version = tracker.snapshot("/path/to/data.json")
data = load(data_path)
modify(data)
tracker.check(version) # raises ConflictError if modified by another process
save(data_path, data)
Locks use os.mkdir() — atomic on all platforms including network filesystems.
Benchmarks
Measured on Apple M4, Python 3.14.
| Memories | Ingest | Search (avg) | Search (p99) | Consolidate | Disk |
|---|---|---|---|---|---|
| 100 | 5.3ms (0.053ms/entry) | 0.40ms | 0.65ms | 4.2ms | 117KB |
| 500 | 16.8ms (0.034ms/entry) | 1.70ms | 2.51ms | 84.3ms | 575KB |
| 1,000 | 33.2ms (0.033ms/entry) | 3.43ms | 5.14ms | 343.3ms | 1.1MB |
| 5,000 | 173.7ms (0.035ms/entry) | 17.10ms | 25.70ms | 4.3s | 5.6MB |
Input gating (P0–P3 classification): 0.177ms avg per input.
Storage Format
workspace/
├── shards/
│ ├── 2026-02-strategic.json
│ ├── 2026-02-operational.json
│ └── 2026-01-tactical.json
├── indexes/
│ ├── search_index.json
│ ├── tag_index.json
│ └── date_index.json
├── migrations/
│ └── history.json
└── memory_audit.json # Deletion audit trail (GDPR)
Plain JSON files. Inspect or edit with any text editor.
Architecture
MemorySystem (v2.0)
├── ShardManager — Date/topic sharding
├── IndexManager — Full-text, tag, and date indexes
│ ├── SearchIndex — BM25 inverted index
│ ├── TagIndex — Tag → hash mapping
│ └── DateIndex — Date range queries
├── SearchEngine — BM25 + optional cosine hybrid
├── MigrationManager — Schema versioning with rollback
├── InputGate — P0-P3 classification at intake
├── DecayEngine — Ebbinghaus forgetting curves
├── ConsolidationEngine — Dedup, clustering, contradiction detection
├── ForgettingEngine — Selective deletion with audit
├── KnowledgeSynthesizer — Gap identification and research integration
├── SharedMemoryPool — Multi-agent coordination
├── NamespaceManager — Multi-tenant isolation
├── ContextPacketBuilder — Sub-agent context injection
├── FileLock — Cross-platform directory-based locking
└── VersionTracker — Optimistic conflict detection
Running Tests
git clone https://github.com/Antaris-Analytics/antaris-memory.git
cd antaris-memory
python -m pytest tests/ -v
All 293 tests pass with zero external dependencies.
Migrating from v1.x
pip install antaris-memory==2.0.0
# Existing workspaces load automatically — no changes required
mem = MemorySystem("./existing_workspace")
mem.load() # Auto-detects format, migrates if needed
# New in v2.0 — opt into typed ingestion
mem.ingest_fact("PostgreSQL supports JSONB")
mem.set_embedding_fn(my_embed_fn) # Enable semantic search
Legacy LegacyMemorySystem (single-file format) still importable: from antaris_memory.core import MemorySystem as LegacyMemorySystem.
Zero Dependencies (Core)
The core package uses only the Python standard library. Optional extras:
pip install mcp— enablescreate_mcp_server()- Supply your own embedding function to
set_embedding_fn()— any callable returninglist[float]works (OpenAI, Ollama, sentence-transformers, etc.)
Why Not a Vector DB?
LangChain Memory, Mem0, and Zep offer embedding-based semantic search and graph relationships. They require cloud APIs or self-hosted infrastructure (Redis, PostgreSQL, vector databases).
Antaris Memory stores everything in plain JSON files, runs fully offline, needs no API keys, and has no vector DB. It uses BM25 keyword ranking by default — good enough for most agent workloads. When you need semantic search, plug in your own embedding function. When you need scale, you already know it.
| Antaris Memory | LangChain Memory | Mem0 | Zep | |
|---|---|---|---|---|
| Search ranking | BM25 + optional cosine | Exact match | Embeddings | Embeddings |
| Input gating | ✅ P0-P3 | ❌ | ❌ | ❌ |
| Memory types | ✅ 5 types | ❌ | ❌ | ❌ |
| Namespace isolation | ✅ | ❌ | ❌ | ⚠️ |
| Context packets | ✅ | ❌ | ❌ | ❌ |
| MCP server | ✅ | ❌ | ❌ | ❌ |
| No database required | ✅ | ❌ | ❌ | ❌ |
| No API keys required | ✅ | ❌ | ❌ | ❌ |
| Memory decay | ✅ Ebbinghaus | ❌ | ❌ | ⚠️ |
| Contradiction detection | ✅ Rule-based | ❌ | ❌ | ⚠️ |
| Selective forgetting | ✅ With audit | ❌ | ⚠️ | ⚠️ |
| Infrastructure needed | None | Redis/PG | Vector + KV + Graph | PostgreSQL + Vector |
Part of the Antaris Analytics Suite
- antaris-memory — Persistent memory for AI agents (this package)
- antaris-router — Adaptive model routing with SLA enforcement
- antaris-guard — Security and prompt injection detection
- antaris-context — Context window optimization
License
Licensed under the Apache License 2.0. See LICENSE for details.
🔗 Related Packages
- antaris-router - Adaptive model routing
- antaris-guard - Security and safety
- antaris-context - Context window optimization
- antaris-pipeline - Agent orchestration pipeline
📞 Support
- Documentation: docs.antarisanalytics.ai
- Email: dev@antarisanalytics.com
- Website: antarisanalytics.ai
Built with ❤️ by Antaris Analytics
Deterministic infrastructure for AI agents
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file antaris_memory-2.1.0.tar.gz.
File metadata
- Download URL: antaris_memory-2.1.0.tar.gz
- Upload date:
- Size: 119.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
d7de9eeed4df6971bac484b91167e6b4bf66d41cd83fb9b8c488a077a463c22e
|
|
| MD5 |
1286905681276c6bf09687f879660d09
|
|
| BLAKE2b-256 |
2efb34a39f6db3c6ecb85053a2c643691848437af45479b23b6f9a42bae615c5
|
File details
Details for the file antaris_memory-2.1.0-py3-none-any.whl.
File metadata
- Download URL: antaris_memory-2.1.0-py3-none-any.whl
- Upload date:
- Size: 91.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
214801b8fbab51cd352555af04bf163870895af3e9820a62b6c1bf42dbd4ddbe
|
|
| MD5 |
eb5d732c5e72aa5aba7152bfcc85e526
|
|
| BLAKE2b-256 |
e0998db97841e61d6be4166b7637430953d2eb02184ca3d767e1c7513c685e3a
|