Skip to main content

Repo memory for AI agents — every memory has a source, every source gets verified.

Project description

agentic-memory (memcite)

PyPI CI License: MIT

pip install memcitefrom agentic_memory import Memory → CLI: am

Let your AI agent remember project settings — and know when they've changed.

The problem

Your AI agent remembers "this project uses Jest for testing." Two weeks later, someone switches to Vitest. The agent doesn't know. It keeps writing Jest tests and breaks your CI.

This isn't hallucination — the memory was correct. It's stale memory, and it's worse than hallucination because the agent is confident about it.

The fix

memcite forces every memory to cite its source. Before using a memory, it checks: is the source still the same?

# Tell the agent "we use ruff" and point to the proof
am add "Uses ruff for linting, line-length=120" --file pyproject.toml --lines 15-20

# Later, ask what linter we use
am query "linting"
# → ✓ Uses ruff for linting, line-length=120
#     pyproject.toml L15-20 [valid]

# Now go change pyproject.toml, then:
am validate
# → ⚠ 1 memory stale (evidence changed)
#     "Uses ruff for linting" ← pyproject.toml L15-20 changed

That's it. Memory with a source. Source gets checked. Stale = flagged.

Python SDK

from agentic_memory import Memory, FileRef

mem = Memory("./my-project")

mem.add(
    "This project uses ruff for linting with line-length=120",
    evidence=FileRef("pyproject.toml", lines=(15, 20)),
)

result = mem.query("What linter does this project use?")
print(result.memories[0].content)           # "ruff with line-length=120"
print(result.citations[0].status.value)     # "valid" or "stale"

Agentic Features (v0.6)

from agentic_memory import Memory, ManualRef

mem = Memory("./my-project")

# Classify memories by kind
mem.add("Never use unsafe patterns", evidence=ManualRef("security review"), kind="antipattern", importance=3)
mem.add("Team prefers early returns", evidence=ManualRef("retro"), kind="preference")

# Query with filters
rules = mem.query("coding standards", kind="rule", min_importance=2)

# Ephemeral memories with TTL (auto-expire after 1 hour)
mem.add("Deploy freeze until 5pm", evidence=ManualRef("slack"), ttl_seconds=3600)

# Deduplication — adding the same content returns the existing record
r1 = mem.add("Uses ruff", evidence=ManualRef("docs"))
r2 = mem.add("Uses ruff", evidence=ManualRef("other"))
assert r1.id == r2.id  # no duplicate

# Retrieval stats
stats = mem.retrieval_stats()
print(f"Last query: {stats[0].query}, latency: {stats[0].latency_ms:.0f}ms")
# CLI: add with kind + importance + TTL
am add "No force push to main" --note "team rule" --kind rule --importance 3
am add "Sprint ends Friday" --note "standup" --ttl 604800  # 1 week

# CLI: query with filters
am query "coding rules" --kind rule --min-importance 2

Phase 2: Agent-Ready Features (v0.7)

from agentic_memory import Memory, ManualRef

mem = Memory("./my-project")

# Conflict detection — warns when new memory contradicts existing ones
mem.add("project uses ruff for code linting", evidence=ManualRef("docs"))
result = mem.add_with_result("project uses black for code linting", evidence=ManualRef("pr"))
if result.conflicts:
    print(f"Conflicts with {len(result.conflicts)} existing memories!")

# create_if_useful — only store if important enough
added = mem.create_if_useful("minor note", evidence=ManualRef("chat"), importance=0, min_importance=2)
assert added is None  # rejected: below threshold

# compact — clean up expired memories
mem.compact()  # returns CompactResult(expired_removed=3, total_before=10, total_after=7)

# search_context — formatted context string for agent prompts
context = mem.search_context("linting rules", kind="rule", min_importance=2)
# → "Found 2 memories:\n[1] No force push to main\n    ✓ team rule [valid]..."

# eval_metrics — system health at a glance
metrics = mem.eval_metrics()
print(f"Queries: {metrics.total_queries}, Avg latency: {metrics.avg_latency_ms}ms")

Design Principles

  1. No Evidence, No Memoryadd() without a citation raises an error
  2. Validate Before Usequery() re-checks citations by default
  3. Decay What's Stale — confidence drops when evidence changes; invalid memories are deprioritized

Evidence Types

Type What it tracks Validation method
FileRef File path + line range Check file exists, content matches
GitCommitRef Commit SHA + file Verify commit exists in history
URLRef Web URL HTTP HEAD check + content hash
ManualRef Human-provided note No auto-validation (always trusted)

Features

  • Repo-scoped — each repository gets its own memory namespace
  • Local-first — SQLite storage, no external services required
  • Citation-backed — every memory traces back to a verifiable source
  • Auto-validation — stale evidence is detected before it misleads your agent
  • Confidence scoring — memories with invalid citations get deprioritized
  • Memory classification — categorize as fact, rule, antipattern, preference, or decision
  • Importance scoring — prioritize memories 0-3 (low → critical), query results sorted by importance
  • TTL / expiration — set time-to-live on ephemeral memories; expired ones are auto-filtered
  • Deduplication — identical content is detected by hash; no duplicate entries
  • Retrieval logging — every query is logged with returned IDs, count, and latency
  • CLI includedam add, am query, am validate, am status

Installation

Status: Alpha but usable — core features are stable, API may evolve.

pip install memcite

With extras:

pip install memcite[mcp]     # MCP server for Claude Code
pip install memcite[api]     # REST API server (FastAPI)

CLI Usage

# Add a memory with file evidence
am add "Uses pytest for testing" --file tests/conftest.py --lines 1-10

# Query memories
am query "What test framework?"

# Validate all memories
am validate

# Show memory status
am status

MCP Server (Claude Code / Cursor / etc.)

memcite includes a built-in MCP server that runs locally on your machine — no cloud service, no API key, no deployment needed. Claude Code spawns it automatically as a subprocess.

Quick setup: add this to your project's .mcp.json:

{
  "mcpServers": {
    "agentic-memory": {
      "command": "am-mcp",
      "args": ["--repo", "/path/to/your/project"]
    }
  }
}

Or use the one-liner: am claude-setup (auto-generates .mcp.json + adds memory protocol to CLAUDE.md)

Once configured, your AI agent gets these tools: memory_add, memory_query, memory_validate, memory_status, memory_list, memory_delete

REST API

am-server --repo /path/to/repo --port 8080

OpenAPI docs at http://localhost:8080/docs. Endpoints:

Method Path Description
POST /memories Add a memory with evidence
POST /memories/query Hybrid search + citation validation
GET /memories List all memories
GET /memories/{id} Get a specific memory
DELETE /memories/{id} Delete a memory
POST /memories/validate Validate all citations
GET /status Memory status summary

Hybrid Search

When initialized with an embedding provider, queries combine FTS5 full-text search with vector similarity:

from agentic_memory import Memory, TFIDFEmbedding, FileRef

mem = Memory("./my-project", embedding=TFIDFEmbedding())
mem.add("Uses ruff for code formatting", evidence=FileRef("pyproject.toml", lines=(1, 5)))

# Finds the memory even though "linting" != "formatting"
result = mem.query("What linter does this project use?")

Default weights: FTS5 (0.65) + Vector (0.35). Customize per query:

result = mem.query("linting", fts_weight=0.5, vector_weight=0.5)

Admission Control

Filter out low-value memories before they're stored:

from agentic_memory import Memory, HeuristicAdmissionController

mem = Memory("./my-project", admission=HeuristicAdmissionController())
mem.add("ok", evidence=ManualRef("chat"))  # raises ValueError — too vague

Or use LLM-based scoring with any OpenAI-compatible API:

from agentic_memory import LLMAdmissionController

def my_llm(system: str, user: str) -> str:
    # Call your LLM here, return JSON: {"score": 0.0-1.0, "reason": "..."}
    ...

mem = Memory("./my-project", admission=LLMAdmissionController(llm_callable=my_llm))

Real-world Workflows

PR reviewer agent — remember repo conventions and enforce them automatically:

mem.add(
    "Logging must use structlog, not stdlib logging",
    evidence=FileRef("docs/conventions.md", lines=(10, 15)),
)

# In your review pipeline
result = mem.query("What logging library should this project use?")
# → "structlog" with citation pointing to docs/conventions.md

Coding agent — look up project config with verifiable sources:

result = mem.query("What env vars does this service need?")
# → Returns memories citing .env.example with current validation status
# If .env.example was deleted or changed, the memory is flagged as STALE

CI pipeline — catch drifted knowledge before it causes damage:

# Add to your CI workflow
am validate --exit-code  # exits non-zero if any memory is INVALID

Roadmap

  • Core SDK — add / query / validate with citation enforcement
  • CLI tool
  • MCP Server — use with Claude Code and other MCP clients
  • Admission control — LLM-based scoring to filter low-value memories
  • Hybrid search — FTS5 + TF-IDF vector fusion, pluggable embedding providers
  • REST API server — FastAPI with OpenAPI docs
  • Agentic features — kind classification, importance scoring, TTL, dedup, retrieval logging
  • GitHub App / GitLab integration (webhook + comment bot)
  • LangChain / LlamaIndex integration
  • Web dashboard

Compared to

mem0 Zep LangMem agentic-memory
Vector search Yes Yes Yes Yes
Forced citations No No No Yes
Source validation No No No Yes
Staleness detection No No No Yes
Repo-scoped No No No Yes
Memory TTL No No No Yes
Deduplication No Partial No Yes
Retrieval logging No No No Yes
Self-hosted Yes Yes Yes Yes

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

memcite-0.7.1.tar.gz (51.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

memcite-0.7.1-py3-none-any.whl (40.5 kB view details)

Uploaded Python 3

File details

Details for the file memcite-0.7.1.tar.gz.

File metadata

  • Download URL: memcite-0.7.1.tar.gz
  • Upload date:
  • Size: 51.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.12

File hashes

Hashes for memcite-0.7.1.tar.gz
Algorithm Hash digest
SHA256 5d63cf5ffc12590c41252beb6c420f0f762bc2a80432605cd17c1bd20352303e
MD5 a3c395700d3806abab07ac3419193526
BLAKE2b-256 57edb89a8881f273ba04e1523901685c8aeb84b8965e7c3a9db5ea70f257cdb1

See more details on using hashes here.

File details

Details for the file memcite-0.7.1-py3-none-any.whl.

File metadata

  • Download URL: memcite-0.7.1-py3-none-any.whl
  • Upload date:
  • Size: 40.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.12

File hashes

Hashes for memcite-0.7.1-py3-none-any.whl
Algorithm Hash digest
SHA256 722dc890e4c4d6aebaf5f2cfad13e80a66c0944c51a29a5ce51ad51806817671
MD5 38e22ee752dd712d685393a457144c99
BLAKE2b-256 5c0c8d0367c5653ef015eeb4ac67e1498f9a7bdda0626027e0763c77a187d2d5

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page