Skip to main content

Engram-inspired memory MCP server with hot cache and pattern mining

Project description

🧠 Memory MCP

Give your AI assistant a persistent second brain

License: MIT Python 3.10+ MCP 1.0 Claude Code PyPI CI


Stop re-explaining your project every session.

Memory MCP learns what matters and keeps it ready — instant recall for the stuff you use most, semantic search for everything else.


The Problem

Every new chat starts from scratch. You explain your architecture again. You paste the same patterns again. Your context window bloats with repetition.

Other memory solutions help, but they still require tool calls for every lookup — adding latency and eating into Claude's thinking budget.

Memory MCP fixes this with a two-tier architecture:

  1. Hot cache (0ms) — Frequently-used knowledge auto-injected into context before Claude even starts thinking. No tool call needed.
  2. Cold storage (~50ms) — Everything else, searchable by meaning via semantic similarity.

The system learns what you use and promotes it automatically. Your most valuable knowledge becomes instantly available. No manual curation required.

Before & After

😤 Without Memory MCP 🎯 With Memory MCP
"Let me explain our architecture again..." Project facts persist and isolate per repo
Copy-paste the same patterns every session Patterns auto-promoted to instant access
500k+ token context windows Hot cache keeps it lean (~20 items)
Tool call latency on every memory lookup Hot cache: 0ms — already in context
Stale information lingers forever Trust scoring demotes outdated facts
Flat list of disconnected facts Knowledge graph connects related concepts

Install

# Install package
uv tool install hot-memory-mcp   # or: pip install hot-memory-mcp

# Add plugin (recommended)
claude plugins add michael-denyer/memory-mcp

The plugin gives you auto-configured hooks, slash commands, and the Memory Analyst agent. MLX is auto-detected on Apple Silicon.

Manual config (no plugin)

Add to ~/.claude.json:

{
  "mcpServers": {
    "memory": {
      "command": "memory-mcp"
    }
  }
}

See Reference for full configuration options.

Restart Claude Code. The hot cache auto-populates from your project docs.

First run: Embedding model (~90MB) downloads automatically. Takes 30-60 seconds once.

How It Works

flowchart LR
    subgraph LLM["Claude"]
        REQ((Request))
    end

    subgraph Hot["HOT CACHE · 0ms"]
        HC[Session context]
        PM[(Promoted memories)]
    end

    subgraph Cold["COLD STORAGE · ~50ms"]
        VS[(Vector search)]
        KG[(Knowledge graph)]
    end

    REQ -->|"auto-injected"| HC
    HC -.->|"draws from"| PM
    REQ -->|"recall()"| VS
    VS <-->|"related"| KG

The hot cache (~10 items) is injected into every request — it combines recent recalls, predicted next memories, and top promoted items. Promoted memories (~20 items) is the backing store of frequently-used memories. Memories used 3+ times auto-promote; unused ones demote after 14 days.

What Makes It Different

Most memory systems make you pay a tool-call tax on every lookup. Memory MCP's hot cache bypasses this entirely — your most-used knowledge is already in context when Claude starts thinking.

Memory MCP Generic Memory Servers
Hot cache Auto-injected at 0ms Every lookup = tool call
Self-organizing Learns and promotes automatically Manual curation required
Project-aware Auto-isolates by git repo One big pile of memories
Knowledge graph Multi-hop recall across concepts Flat list of facts
Pattern mining Learns from Claude's outputs Not available
Trust scoring Outdated info decays and sinks All memories equal
Setup One command, local SQLite Often needs cloud setup

The Engram Insight: Human memory doesn't search — frequently-used patterns are already there. That's what hot cache does for Claude.

Quick Reference

Slash Command Tool Description
/memory-mcp:remember remember Store a memory with semantic embedding
/memory-mcp:recall recall Search memories by meaning
/memory-mcp:hot-cache promote / demote Manage promoted memories
/memory-mcp:stats memory_stats Show statistics
/memory-mcp:bootstrap bootstrap_project Seed from project docs
link_memories Knowledge graph connections

See Reference for all 14 slash commands and full tool API.

Dashboard

memory-mcp-cli dashboard    # Opens at http://localhost:8765

Dashboard

Browse memories, hot cache, mining candidates, sessions, and knowledge graph.

How to Use

Memory MCP is designed to run as three complementary components:

Component Purpose
Claude Code Plugin Hooks, slash commands, and Memory Analyst agent for seamless integration
MCP Server Core memory tools available to Claude via Model Context Protocol
Dashboard Web UI to browse, manage, and debug your memory database

The plugin is recommended for most users — it auto-configures the MCP server and adds productivity features. Run the dashboard alongside when you want visibility into what's being stored.

Documentation

Document Description
Reference Full API, CLI, configuration, MCP resources
Troubleshooting Common issues and solutions

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

hot_memory_mcp-0.7.3.tar.gz (686.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

hot_memory_mcp-0.7.3-py3-none-any.whl (197.3 kB view details)

Uploaded Python 3

File details

Details for the file hot_memory_mcp-0.7.3.tar.gz.

File metadata

  • Download URL: hot_memory_mcp-0.7.3.tar.gz
  • Upload date:
  • Size: 686.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for hot_memory_mcp-0.7.3.tar.gz
Algorithm Hash digest
SHA256 0ee8f075cd0baa50ee8cfd7551a9ea7d1901bf7816ff53cb8ce7579ac7dbf47f
MD5 0c9ed9a0781571bc45687c99ab8b8c18
BLAKE2b-256 f8a2a851affe0ea9b048081873d8b6745cc3fd4bffd4d090be4885406bfe94e5

See more details on using hashes here.

Provenance

The following attestation bundles were made for hot_memory_mcp-0.7.3.tar.gz:

Publisher: publish.yml on michael-denyer/memory-mcp

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file hot_memory_mcp-0.7.3-py3-none-any.whl.

File metadata

  • Download URL: hot_memory_mcp-0.7.3-py3-none-any.whl
  • Upload date:
  • Size: 197.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for hot_memory_mcp-0.7.3-py3-none-any.whl
Algorithm Hash digest
SHA256 6c21cf1309c43e1ded7d182dc1c8755ff05c7af7b713e2904b9aca5723b1209a
MD5 21d401c65a917aa44b9704ecb578aad9
BLAKE2b-256 2845fdf545c3346a2647f23f4a93b9a74bd89e532bb6e834532c9f2bd4b40935

See more details on using hashes here.

Provenance

The following attestation bundles were made for hot_memory_mcp-0.7.3-py3-none-any.whl:

Publisher: publish.yml on michael-denyer/memory-mcp

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page