Open-source agent-memory MCP server. recall.works
Project description
Recall™
Memory + coordination for multiple AI agents working on the same codebase. MCP-native. Self-hosted. One Docker image.
Why coordination? · OSS quickstart · Recall Pro → · Book a demo · IceWhisperer for Encompass
Most memory servers (mem0, Letta, Zep) optimize for one agent across many sessions. Recall optimizes for many agents in one session — the actual problem when you run two Copilot windows or two Claude instances on the same monorepo and they clobber each other.
Recall gives them shared memory and soft locks, handoffs, and a live view of what the other agents are doing right now.
┌──────────────┐ ┌──────────────┐
│ Agent a3f7 │ claim(file, ttl) │ Agent b1c4 │
│ Claude #1 │ ───────────┐ ┌─────────► │ Claude #2 │
└──────┬───────┘ ▼ │ └──────┬───────┘
│ ┌────────┴───────┐ │
│ remember │ Recall │ pulse │
├────────────► │ • shared memory│ ◄────────┤
│ │ • claims/locks │ │
│ handoff │ • handoffs │ handoff │
├────────────► │ • who_has │ ◄────────┤
│ └────────────────┘ │
▼ ▼
22 MCP tools — Copilot, Claude, Cursor, custom
Why coordination? (The wedge)
If you run a single AI agent, you don't have the problem Recall solves. The problem starts the moment you run two:
- Two Copilot chat windows on the same repo, both editing
auth/ - A planner agent + an executor agent on the same task
- Three Claude instances dividing up a refactor across folders
- A
pre-commitagent and acode-reviewagent racing on the same PR
No memory tool today addresses this. They all assume one agent. Recall adds six MCP primitives that turn parallel agents from a clobber-fest into a coordinated team:
| Tool | What it does |
|---|---|
claim(resource, agent) |
Soft-lock a file/table/URL with an auto-expiring TTL |
release(resource, agent) |
Drop the lock (soft-archive — audit trail survives) |
who_has(resource) |
"Is anyone editing src/foo.py right now?" |
claims() |
All active locks across all agents |
handoff(to_agent, ...) |
Explicit work transfer with intent + files + context |
pulse_others(self_agent) |
The N most recent checkpoints from agents other than you |
Claims are advisory (like git locks) — Recall doesn't physically stop a second agent from writing, but every well-behaved client checks first. TTLs prevent a crashed agent from freezing a resource forever. Releases soft-archive (per the project-wide delete=archive rule) so the audit trail of who held what when survives.
Plus everything you'd expect from a memory server: 16 more tools for remember, recall, vector + filtered search, checkpoint, reflect, anti-pattern, snapshot, reindex, and stats. 22 MCP tools total.
How is this different from mem0 / Letta / Zep?
They're built for one agent across sessions ("remember what the user said last week"). Recall is built for multiple agents in one session ("don't let agent B overwrite the function agent A is mid-refactoring"). Different problem. Different primitives. Use both — they don't conflict.
One-line install (Claude Desktop, VS Code, Cursor)
Recall ships as a stdio MCP server. Zero config — no API keys, no Docker, no
ports. Memory lives in ~/.recall/.
pip install "ai-recallworks[mcp]"
Then add Recall to your MCP client config:
Claude Desktop (~/Library/Application Support/Claude/claude_desktop_config.json
on macOS, %APPDATA%\Claude\claude_desktop_config.json on Windows):
{
"mcpServers": {
"recall": {
"command": "recall-mcp"
}
}
}
VS Code (mcp.json in your workspace or user settings):
{
"servers": {
"recall": {
"command": "recall-mcp"
}
}
}
Restart the client. Your agent now has persistent memory across sessions. Embeddings run fully offline (Chroma's bundled all-MiniLM-L6-v2). Upgrade to Ollama / OpenAI / Voyage embeddings via env vars when you want.
Five-minute install (HTTP / multi-user / team)
1. Run the server:
docker run -d --name recall \
-p 8787:8787 \
-e API_KEY=changeme \
-v recall-data:/data \
ghcr.io/recallworks/recall:latest
2. Talk to it — pick your stack:
# Raw HTTP (any language)
curl -H "X-API-Key: changeme" \
-H "Content-Type: application/json" \
-d '{"content":"first memory","tags":"hello"}' \
http://localhost:8787/tool/remember
# Python (use requests/httpx — no SDK pkg needed)
import requests
h = {"X-API-Key": "changeme", "Content-Type": "application/json"}
requests.post("http://localhost:8787/tool/remember", headers=h,
json={"content": "first memory", "tags": "hello"})
print(requests.post("http://localhost:8787/tool/recall", headers=h,
json={"query": "memory"}).json()["result"])
// TypeScript / JavaScript (Node 18+, Bun, Deno, browser)
npm install @recallworks/recall-client
import { RecallClient } from "@recallworks/recall-client";
const c = new RecallClient({ baseUrl: "http://localhost:8787", apiKey: "changeme" });
await c.remember("first memory", { tags: "hello" });
console.log((await c.recall("memory")).result);
Full walkthrough: docs/quickstart.md.
What you get
- 13 tools —
remember,recall,reflect,anti_pattern,checkpoint,pulse,session_close,index_file,reindex,snapshot_index,memory_stats,forget,maintenance. - Two transports — plain HTTP (
POST /tool/{name}) and MCP over SSE. Drop into Copilot, Claude Code, Cursor, or any MCP client. - Bring your own models — pluggable embedder (default / OpenAI / Ollama) and summarizer (noop / OpenAI / Ollama). Run fully offline, fully on-prem, or against your own Azure-OpenAI tenant. See docs/byo-models.md.
- Durable by default — ephemeral live store with auto-snapshot to disk; container restarts come up whole.
- Append-only artifacts — every write also lands as a
.mdfile. If the vector store ever burns down,reindexrebuilds it from the artifacts. forgetis soft-archive — guardrail wired into the OSS code itself, not bolted on as policy. Memory you delete can be recovered.
How it's different
| Recall | Mem0 / Letta / Zep | |
|---|---|---|
| License (core) | MIT | mixed; SaaS-first |
| Self-host | one docker run |
varies, often non-trivial |
| BYO embedder | default / OpenAI / Ollama (env var) | usually fixed |
| BYO LLM | noop / OpenAI / Ollama (env var) | usually fixed |
| Storage model | append-only artifacts + vector index, rebuildable | live DB only |
delete |
soft-archive by design | hard delete |
| Tool surface | 13 opinionated tools (memory + workflow) | embedding + retrieval primitives |
| MCP-native | yes, plus plain HTTP | partial / via wrapper |
| Ops model | single binary, single container | multi-service stack |
If you want a managed service, see Recall Cloud below. If you want a brain you fully own, this OSS core is enough.
Repo layout
| Path | What |
|---|---|
src/recall/ |
OSS server (MIT) |
src/recall/tools/ |
One module per tool |
src/recall/transport/ |
HTTP + MCP/SSE adapters |
docker/single-tenant/ |
Reference Dockerfile + compose |
tests/ |
pytest suite (no Docker required) |
docs/ |
Quickstart, conventions, architecture |
enterprise/ |
Multi-tenant, SSO, control plane (BSL) |
Conventions
These are the practices that make the tools pay off. Pick what fits.
- Cold-start ritual — opening protocol every session should run.
- Branding — signed-edit headers so you can trace which agent touched which file when.
Status
Alpha. The code in src/recall/ is extracted from a hosted production brain
that has served thousands of sessions, then sanitized of org-specific
paths, extensions, and tenant data. Expect breaking changes before 1.0; pin
the image tag.
Contributing
Yes — please read CONTRIBUTING.md first. We accept bug
fixes, new Store backends, doc improvements, and anti-pattern entries. We
don't accept architectural rewrites without prior discussion.
Security issues: see SECURITY.md.
License
src/recall/,clients/,docker/single-tenant/,docs/,examples/— MIT (LICENSE)enterprise/— BSL 1.1, 5-seat additional-use grant, converts to MIT after 3 years (LICENSE-COMMERCIAL.md)
Recall Open Source vs. Recall Pro vs. Hosted
| Capability | OSS (this repo) | Recall Pro | Recall Cloud |
|---|---|---|---|
| Single-tenant Docker image | ✅ | ✅ | n/a (hosted) |
| 13 memory tools, MCP + HTTP | ✅ | ✅ | ✅ |
| BYO embedder + LLM | ✅ | ✅ | ✅ |
| Append-only artifacts + auto-snapshot | ✅ | ✅ | ✅ |
| Multi-tenant, SSO, RBAC | — | ✅ | ✅ |
| Audit log + retention policy | — | ✅ | ✅ |
| Cross-session entity graph | — | ✅ | ✅ |
| PII sanitization pipeline | — | ✅ | ✅ |
| Snapshot replication / DR | — | ✅ | ✅ |
| Vendor support + SLA | community | business hours | 24×7 |
| Hosted on our infra | — | — | ✅ |
| Pricing | free | from $99/mo per node | from $0.10 per 1k tools |
Recall Pro ships from the enterprise/ tree under a Business Source License — source-available, 5-seat free Additional Use Grant, converts to MIT after 3 years. Buy a license and the enterprise/ modules light up alongside your OSS install.
Recall Cloud is the hosted multi-tenant version. Same tools, no infra. Reach out for early-access pricing.
➡️ Talk to sales: sales@recall.works · Book a 20-min walkthrough: https://recall.works/demo
Vertical builds powered by Recall
Recall is the engine. We ship turn-key vertical brains on top of it:
- IceWhisperer — the memory + workflow brain for ICE Mortgage Technology / Encompass shops. Pre-loaded SDK index, settings recipes, plugin audits, drift detection. Pilots from $250/mo.
If you want a vertical brain for your industry, we'll build it. Email partners@recall.works.
Maintainers
Reach the maintainers at maintainers@recall.works. Issues and PRs welcome on GitHub.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file ai_recallworks-0.5.0.tar.gz.
File metadata
- Download URL: ai_recallworks-0.5.0.tar.gz
- Upload date:
- Size: 78.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.10
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
ddbc92bcde8433f56b8675d3ab379b24e6e915d50031987148264c662315d2e3
|
|
| MD5 |
612d427c4367b3e3483f335363bb5180
|
|
| BLAKE2b-256 |
9c6ad645a07b3b94e74513f436fa9bc9514bc833c144d3df5468991b94fbe76d
|
File details
Details for the file ai_recallworks-0.5.0-py3-none-any.whl.
File metadata
- Download URL: ai_recallworks-0.5.0-py3-none-any.whl
- Upload date:
- Size: 60.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.10
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
0987e61874143e001978faf73564490bf2cc45d7316dc93080e691d4e6a6e6ed
|
|
| MD5 |
73fba28d8a39733eb963161a797c7bde
|
|
| BLAKE2b-256 |
8b2a6fa46f6ddc2fc62de2893e2eaec0414332ae531bc81f398ebf89a6e87f79
|