Skip to main content

MCP Server that brings long-term memory to AI coding tools via EverMemOS

Project description

evermemos-mcp

PyPI Python CI Smithery License: MIT

English | 简体中文

Long-term memory for AI coding assistants. Remember once, recall forever.

You spent thirty minutes explaining your architecture, naming conventions, and why you dropped MongoDB. Next session — gone. You explain it all over again.

evermemos-mcp fixes this. One remember call stores it. One briefing call brings it back — across any session, any client.

Benchmark: 60/60 recall vs 0/60 baseline. Zero attribution errors. P95 < 2s. (evidence)


Quick Start

Get your API key from EverMemOS Cloud, then add to your MCP client config:

{
  "mcpServers": {
    "evermemos-mcp": {
      "type": "stdio",
      "command": "uvx",
      "args": ["evermemos-mcp@latest"],
      "env": {
        "EVERMEMOS_API_KEY": "your-key-here"
      }
    }
  }
}

Or run directly:

uvx evermemos-mcp@latest

Works with Claude Code, Cursor, Cline, Cherry Studio, and any MCP-compatible client. See docs/05-client-integrations.md for client-specific setup.

Install from source
git clone https://github.com/tt-a1i/evermemos-mcp.git
cd evermemos-mcp
cp .env.example .env   # set EVERMEMOS_API_KEY
uv run evermemos-mcp

MCP client config for source installs:

{
  "mcpServers": {
    "evermemos-mcp": {
      "type": "stdio",
      "command": "uv",
      "args": ["run", "--directory", "/path/to/evermemos-mcp", "evermemos-mcp"],
      "env": { "EVERMEMOS_API_KEY": "your-key-here" }
    }
  }
}

What You Get

7 Tools

Tool What it does
list_spaces Discover available memory spaces
remember Store context into long-term memory. Auto-detects sensitive content (API keys, passwords) and checks for conflicting memories
request_status Check if a queued write has been extracted
recall Search memories with 6 retrieval strategies (keyword / hybrid / vector / RRF / agentic / auto)
briefing One-call session-start context restore: profile + episodes + facts + foresights
forget Targeted deletion with verification workflow
fetch_history Paginate through memory timeline by type

Key Capabilities

  • Space isolationcoding:my-app, chat:preferences, study:ml-notes — memories never bleed across projects
  • Multi-space search — Query up to 10 spaces in one recall call with automatic source attribution
  • Sensitive content guard — Blocks API keys, passwords, tokens, private keys before storing. Asks user to confirm
  • Memory conflict detection — Auto-checks for similar memories in chat:* spaces. Surfaces conflicts so the agent can decide
  • Lifecycle tracking — Every result labeled queued, provisional, fallback, or searchable across all tools
  • Traceable citationsmemory_type, snippet, timestamp, score, source_message_id on every result
  • Git auto-detection — Omit space_id and it infers coding:<repo-name> from git remote
  • Robust error handling — Retry with backoff (429/5xx), GET body fallback for proxy/WAF, structured error codes

Use Cases

Persistent architecture context:

You: remember we chose PostgreSQL because our data is highly relational
     [space_id: coding:my-saas]

-- next day, new session --

You: what database did we choose and why?
     → "Chose PostgreSQL — highly relational data model"

Personal preferences that stick:

You: remember I prefer dark mode, vim keybindings, and concise responses
     [space_id: chat:preferences]

-- any future session --

You: recall my UI preferences
     → "dark mode, vim keybindings, concise responses"

Cross-session learning notes:

You: remember bias-variance tradeoff — high bias = underfitting, high variance = overfitting
     [space_id: study:ml-notes]

-- later --

You: briefing for study:ml-notes
     → profile + recent episodes + key facts + foresights

Why evermemos-mcp

There are other memory MCP servers. Here's what makes this one different:

evermemos-mcp Mem0 MCP Letta/MemGPT Official MCP memory
Space isolation domain:slug per project/topic No No No
Lifecycle tracking queued → provisional → fallback → searchable No No No
Sensitive content guard API keys, passwords, tokens blocked No No No
Conflict detection Auto for chat spaces No No No
Multi-space search Up to 10 spaces in one call No No No
Retrieval strategies 6 methods + auto merge Semantic only Semantic only None
Benchmark verified 60/60 recall, 0 errors
Setup uvx evermemos-mcp Cloud or self-host Self-host required npx

Benchmark

Tested on a fixed 60-query set across coding, chat, and study spaces.

Metric With memory Without memory
Hit rate 60/60 (100%) 0/60 (0%)
Attribution errors 0
P95 latency 1958 ms

Evidence:


How It Works

MCP Client (Claude Code / Cursor / Cline / Cherry Studio)
        │
        │  MCP stdio
        ▼
┌─────────────────────────────┐
│     evermemos-mcp server    │
│  ┌───────────────────────┐  │
│  │   7 Tool Handlers     │  │
│  └──────────┬────────────┘  │
│  ┌──────────▼────────────┐  │
│  │   Memory Service      │  │  Content guard → Conflict check → Cloud write → Lifecycle tracking
│  └──────────┬────────────┘  │
│  ┌──────────▼────────────┐  │
│  │ Space Catalog Service │  │  Space registry, metadata sync, cross-session recovery
│  └──────────┬────────────┘  │
│  ┌──────────▼────────────┐  │
│  │  EverMemOS HTTP Client│  │  Auth, retries, rate-limit backoff, error normalization
│  └──────────┬────────────┘  │
└─────────────┼───────────────┘
              │  HTTPS
              ▼
       EverMemOS Cloud API
  • Cloud-first — All memories live in EverMemOS Cloud. No local state to lose.
  • Async extractionremember queues content for AI extraction. Use request_status to track progress.
  • Not a thin wrapper — 2500+ lines of orchestration: fallback hierarchies, multi-method search merging, identity mirroring, partial failure recovery.

Space Templates

Template Use it for
chat:preferences Durable personal preferences, names, tone, UI likes
chat:daily Ongoing chat context that shouldn't leak into projects
coding:<repo> Architecture decisions, conventions, bugs, project context
study:<topic> Learning notes, topic progress, revision context

Which Tool When

Goal Tool Why
Start a new session briefing Fastest way to restore context in one call
Find a specific fact recall Relevance-ranked search across spaces
Review what happened fetch_history Chronological timeline > ranked search for audits
Verify before/after delete fetch_history Stable timeline for pre/post-delete checks

Configuration

Variable Default Description
EVERMEMOS_API_KEY (required) EverMemOS Cloud API key
EVERMEMOS_USER_ID mcp-user Default user identity
EVERMEMOS_DEFAULT_SPACE (auto) Default space. Auto-detected from git remote as coding:<repo>
EVERMEMOS_BASE_URL https://api.evermind.ai API endpoint
EVERMEMOS_DEFAULT_TIMEZONE UTC Timezone for metadata
EVERMEMOS_ENABLE_CONVERSATION_META true Sync conversation metadata
Advanced configuration
Variable Default Description
EVERMEMOS_API_VERSION v0 API version
EVERMEMOS_LLM_CUSTOM_SETTING_JSON Custom LLM extraction settings
EVERMEMOS_USER_DETAILS_JSON User profile details for conversations

flush Rules

Scenario flush
Mid-conversation, more messages coming false
End of session / topic switch / summary true
Uncertain true (safer)

Advanced: Memory Lifecycle States
State Meaning
queued Write accepted, extraction not yet confirmed
provisional Answer from pending_messages while extraction is in progress
fallback Answer from mirrored conversation-meta, not formal extracted memory
searchable Answer from formal extracted memories

All 7 tools expose compatible lifecycle blocks so agents always know memory maturity.

Advanced: Forget Safety

Cloud deletion is async and best-effort. evermemos-mcp provides a verification-first workflow:

  1. Confirm target memory_id via fetch_history or recall
  2. Call forget(memory_ids=[...], space_id=...)
  3. Verify with fetch_history
  4. If target persists, the lifecycle model surfaces this transparently

This is deliberate: expose real state to the agent rather than pretend deletion is instant.


Development

uv sync --group dev       # Install dev dependencies
uv run ruff check         # Lint
uv run pytest             # Tests (285 pass)

Documentation

Document Description
docs/02-architecture.md Technical architecture
docs/05-client-integrations.md Client setup guides
docs/auto-memory-prompt.md Auto-memory prompt templates
docs/06-benchmark.md Benchmark protocol
CHANGELOG.md Version history

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

evermemos_mcp-0.5.2.tar.gz (212.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

evermemos_mcp-0.5.2-py3-none-any.whl (48.9 kB view details)

Uploaded Python 3

File details

Details for the file evermemos_mcp-0.5.2.tar.gz.

File metadata

  • Download URL: evermemos_mcp-0.5.2.tar.gz
  • Upload date:
  • Size: 212.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for evermemos_mcp-0.5.2.tar.gz
Algorithm Hash digest
SHA256 60e3974747a384ffddc4762eeffb77f804cf7434628f64e5aefdc2db1ae7bb03
MD5 d5091ee3cf456cc11549a155a61b9635
BLAKE2b-256 cf7f53d50ceb04256ba15606ced881770349b6eff88602ed81a2232e21829fe9

See more details on using hashes here.

Provenance

The following attestation bundles were made for evermemos_mcp-0.5.2.tar.gz:

Publisher: publish.yml on tt-a1i/evermemos-mcp

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file evermemos_mcp-0.5.2-py3-none-any.whl.

File metadata

  • Download URL: evermemos_mcp-0.5.2-py3-none-any.whl
  • Upload date:
  • Size: 48.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for evermemos_mcp-0.5.2-py3-none-any.whl
Algorithm Hash digest
SHA256 82ce55b295a7e82eae2652daf77e9f99efd27b81ce8f00e5574fa9d3e29a1634
MD5 db90e231fb877173667b92077e8fb1fb
BLAKE2b-256 2ccfd73a81aefd0c495003b8ed63264e8190193122f33ecfb6e3d20c91ea9a3a

See more details on using hashes here.

Provenance

The following attestation bundles were made for evermemos_mcp-0.5.2-py3-none-any.whl:

Publisher: publish.yml on tt-a1i/evermemos-mcp

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page