Skip to main content

Persistent context for AI coding agents. No cloud. Just markdown.

Project description

poma-memory

Your AI agent loses context every session. poma-memory gives it back.

AI coding agents (Claude Code, Cursor, Copilot) accumulate valuable project knowledge — decisions, architecture notes, task history — in markdown files (.claude/, .cursor/, .github/copilot/). But when context windows fill up or sessions restart, that knowledge becomes invisible. Grep finds strings; it doesn't understand structure.

poma-memory indexes those markdown files and returns complete, readable context — not disconnected snippets. When you search for "auth middleware", you get the matching paragraph plus its parent headings and surrounding context, assembled into a coherent cheatsheet with [...] gap markers. The result reads like a compressed version of the original document.

No POMA account or API key required. Free, local, open-source.


The Problem

Standard search (grep, embeddings over flat chunks) breaks document structure:

  • Orphaned content — a paragraph arrives without its heading, so the agent doesn't know what section it belongs to
  • Lost hierarchy — a nested list item loses its parent context
  • Fragment soup — five hits from the same file come back as five disconnected blocks instead of one readable summary

For agent memory files — which are deeply hierarchical by design — this means the agent retrieves text but not understanding.

The Solution

poma-memory preserves the full document hierarchy during chunking. Every retrieval unit is a root-to-leaf path through the heading tree, so results always carry complete context. Multiple hits from the same file are merged into a single cheatsheet.

This is an open-source extraction of POMA's heuristic chunking engine, optimized for the clean, consistent markdown that AI agents produce.


Install

pip install poma-memory                         # BM25 keyword search (always works)
pip install poma-memory[semantic]               # + model2vec local embeddings (30MB, no API key)
pip install poma-memory[openai]                 # + OpenAI text-embedding-3-large
pip install poma-memory[mcp]                    # + MCP server for Claude Code
pip install poma-memory[semantic,mcp]           # recommended combo

Quick start

poma-memory index .claude/                                    # index your context files
poma-memory search "authentication middleware" --path .claude/ # search

MCP server (Claude Code)

Add poma-memory as an MCP server so Claude Code can search your project memory automatically:

claude mcp add --transport stdio --scope user poma-memory -- poma-memory-mcp
# Exposes poma_search, poma_index, poma_status tools

Once added, Claude Code can call poma_search during planning and exploration to surface relevant decisions, patterns, and context from prior sessions.

Python API

from poma_memory import index, search, status

index(path=".claude/")
results = search("session context", path=".claude/", top_k=5)
for r in results:
    print(f"{r['file_path']} (score: {r['score']:.4f})")
    print(r['context'])

How it works

  1. Hierarchical chunking. Markdown is parsed into depth-annotated chunks that preserve heading hierarchy, lists, code blocks, and tables.
  2. Chunkset assembly. Leaf chunks are paired with their ancestors into self-contained retrieval units (root-to-leaf paths), so every result carries full context.
  3. Hybrid search. BM25 keyword matching (always available) + optional semantic vectors, merged via Reciprocal Rank Fusion.
  4. Cheatsheet merging. Multiple hits from the same file are merged into one block with [...] gap markers — reads like a summary, not a list of excerpts.
  5. Incremental indexing. Append-only files (like agent context logs) only process new content on re-index.

Search backends

Backend Install Requires Best for
BM25 included nothing exact term matching
model2vec [semantic] 30MB local model general-purpose, no API key
OpenAI [openai] OPENAI_API_KEY highest quality

Multiple backends are automatically fused via Reciprocal Rank Fusion when available.


What this is (and isn't)

poma-memory extracts the heuristic chunking and retrieval logic from POMA's document processing platform. It works well on clean, predictable markdown — exactly what agents produce.

It does not include POMA's ML-powered indentation analysis, fine-tuned embedding models, or cloud processing pipeline. For complex document processing (scanned PDFs, inconsistent formatting, enterprise scale), see poma-ai.com.

Built for

  • AI coding agents that persist context in markdown (.claude/, .cursor/, .github/copilot/)
  • Megavibe multi-agent framework (ships with poma-memory integration)
  • Claude Code hook pipelines (augment Grep results with semantic context)

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

poma_memory-0.3.3.tar.gz (21.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

poma_memory-0.3.3-py3-none-any.whl (18.8 kB view details)

Uploaded Python 3

File details

Details for the file poma_memory-0.3.3.tar.gz.

File metadata

  • Download URL: poma_memory-0.3.3.tar.gz
  • Upload date:
  • Size: 21.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for poma_memory-0.3.3.tar.gz
Algorithm Hash digest
SHA256 d7a9765eada7e81ef74c4ba45fe881ccd58ba2b2c40ab13e47415c76fd45c60e
MD5 bab9851ef1aff1a2ec779f85d7e8b4b8
BLAKE2b-256 e97943c6fa4dad0366c4504db93951b11d601547b355687b113ec8acd5bad653

See more details on using hashes here.

Provenance

The following attestation bundles were made for poma_memory-0.3.3.tar.gz:

Publisher: pypi-publish.yml on poma-ai/poma-memory

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file poma_memory-0.3.3-py3-none-any.whl.

File metadata

  • Download URL: poma_memory-0.3.3-py3-none-any.whl
  • Upload date:
  • Size: 18.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for poma_memory-0.3.3-py3-none-any.whl
Algorithm Hash digest
SHA256 f4208cab24253dff0b808cc5681cb901111f54151c81bd16f367ebf0413d00ba
MD5 bb58ea387530ed04a7ab9549f8cbb238
BLAKE2b-256 93aab8a2ba1450faf643d1bb7fef3fb45a097d8b01c16f70e6c4b36fc4b5f095

See more details on using hashes here.

Provenance

The following attestation bundles were made for poma_memory-0.3.3-py3-none-any.whl:

Publisher: pypi-publish.yml on poma-ai/poma-memory

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page