Skip to main content

Semantic memory search for markdown knowledge bases

Project description

ย  memsearch

OpenClaw's memory, everywhere.

PyPI Claude Code Plugin Python License Docs Stars Discord X (Twitter)

https://github.com/user-attachments/assets/31de76cc-81a8-4462-a47d-bd9c394d33e3

๐Ÿ’ก Give your AI agents persistent memory in a few lines of code. Write memories as markdown, search them semantically. Inspired by OpenClaw's markdown-first memory architecture. Pluggable into any agent framework.

โœจ Why memsearch?

  • ๐Ÿ“ Markdown is the source of truth โ€” human-readable, git-friendly, zero vendor lock-in. Your memories are just .md files
  • โšก Smart dedup โ€” SHA-256 content hashing means unchanged content is never re-embedded
  • ๐Ÿ”„ Live sync โ€” File watcher auto-indexes changes to the vector DB, deletes stale chunks when files are removed
  • ๐Ÿงฉ Ready-made Claude Code plugin โ€” a drop-in example of agent memory built on memsearch

๐Ÿ“ฆ Installation

pip install memsearch
Optional embedding providers
pip install "memsearch[google]"      # Google Gemini
pip install "memsearch[voyage]"      # Voyage AI
pip install "memsearch[ollama]"      # Ollama (local)
pip install "memsearch[local]"       # sentence-transformers (local, no API key)
pip install "memsearch[all]"         # Everything

๐Ÿ Python API โ€” Give Your Agent Memory

from memsearch import MemSearch

mem = MemSearch(paths=["./memory"])

await mem.index()                                      # index markdown files
results = await mem.search("Redis config", top_k=3)    # semantic search
print(results[0]["content"], results[0]["score"])       # content + similarity
๐Ÿš€ Full example โ€” agent with memory (OpenAI) โ€” click to expand
import asyncio
from datetime import date
from pathlib import Path
from openai import OpenAI
from memsearch import MemSearch

MEMORY_DIR = "./memory"
llm = OpenAI()                                        # your LLM client
mem = MemSearch(paths=[MEMORY_DIR])                    # memsearch handles the rest

def save_memory(content: str):
    """Append a note to today's memory log (OpenClaw-style daily markdown)."""
    p = Path(MEMORY_DIR) / f"{date.today()}.md"
    p.parent.mkdir(parents=True, exist_ok=True)
    with open(p, "a") as f:
        f.write(f"\n{content}\n")

async def agent_chat(user_input: str) -> str:
    # 1. Recall โ€” search past memories for relevant context
    memories = await mem.search(user_input, top_k=3)
    context = "\n".join(f"- {m['content'][:200]}" for m in memories)

    # 2. Think โ€” call LLM with memory context
    resp = llm.chat.completions.create(
        model="gpt-4o-mini",
        messages=[
            {"role": "system", "content": f"You have these memories:\n{context}"},
            {"role": "user", "content": user_input},
        ],
    )
    answer = resp.choices[0].message.content

    # 3. Remember โ€” save this exchange and index it
    save_memory(f"## {user_input}\n{answer}")
    await mem.index()

    return answer

async def main():
    # Seed some knowledge
    save_memory("## Team\n- Alice: frontend lead\n- Bob: backend lead")
    save_memory("## Decision\nWe chose Redis for caching over Memcached.")
    await mem.index()  # or mem.watch() to auto-index in the background

    # Agent can now recall those memories
    print(await agent_chat("Who is our frontend lead?"))
    print(await agent_chat("What caching solution did we pick?"))

asyncio.run(main())
๐Ÿ’œ Anthropic Claude example โ€” click to expand
pip install memsearch anthropic
import asyncio
from datetime import date
from pathlib import Path
from anthropic import Anthropic
from memsearch import MemSearch

MEMORY_DIR = "./memory"
llm = Anthropic()
mem = MemSearch(paths=[MEMORY_DIR])

def save_memory(content: str):
    p = Path(MEMORY_DIR) / f"{date.today()}.md"
    p.parent.mkdir(parents=True, exist_ok=True)
    with open(p, "a") as f:
        f.write(f"\n{content}\n")

async def agent_chat(user_input: str) -> str:
    # 1. Recall
    memories = await mem.search(user_input, top_k=3)
    context = "\n".join(f"- {m['content'][:200]}" for m in memories)

    # 2. Think โ€” call Claude with memory context
    resp = llm.messages.create(
        model="claude-sonnet-4-5-20250929",
        max_tokens=1024,
        system=f"You have these memories:\n{context}",
        messages=[{"role": "user", "content": user_input}],
    )
    answer = resp.content[0].text

    # 3. Remember
    save_memory(f"## {user_input}\n{answer}")
    await mem.index()
    return answer

async def main():
    save_memory("## Team\n- Alice: frontend lead\n- Bob: backend lead")
    await mem.index()
    print(await agent_chat("Who is our frontend lead?"))

asyncio.run(main())
๐Ÿฆ™ Ollama (fully local, no API key) โ€” click to expand
pip install "memsearch[ollama]"
ollama pull nomic-embed-text          # embedding model
ollama pull llama3.2                  # chat model
import asyncio
from datetime import date
from pathlib import Path
from ollama import chat
from memsearch import MemSearch

MEMORY_DIR = "./memory"
mem = MemSearch(paths=[MEMORY_DIR], embedding_provider="ollama")

def save_memory(content: str):
    p = Path(MEMORY_DIR) / f"{date.today()}.md"
    p.parent.mkdir(parents=True, exist_ok=True)
    with open(p, "a") as f:
        f.write(f"\n{content}\n")

async def agent_chat(user_input: str) -> str:
    # 1. Recall
    memories = await mem.search(user_input, top_k=3)
    context = "\n".join(f"- {m['content'][:200]}" for m in memories)

    # 2. Think โ€” call Ollama locally
    resp = chat(
        model="llama3.2",
        messages=[
            {"role": "system", "content": f"You have these memories:\n{context}"},
            {"role": "user", "content": user_input},
        ],
    )
    answer = resp.message.content

    # 3. Remember
    save_memory(f"## {user_input}\n{answer}")
    await mem.index()
    return answer

async def main():
    save_memory("## Team\n- Alice: frontend lead\n- Bob: backend lead")
    await mem.index()
    print(await agent_chat("Who is our frontend lead?"))

asyncio.run(main())

๐Ÿ“– Full Python API reference with all parameters โ†’ Python API docs

๐Ÿ–ฅ๏ธ CLI Usage

Set Up โ€” config init

Interactive wizard to configure embedding provider, Milvus backend, and chunking parameters:

memsearch config init                    # write to ~/.memsearch/config.toml
memsearch config init --project          # write to .memsearch.toml (per-project)
memsearch config set milvus.uri http://localhost:19530
memsearch config list --resolved         # show merged config from all sources

Index Markdown โ€” index

Scan directories and embed all markdown into the vector store. Unchanged chunks are auto-skipped via content-hash dedup:

memsearch index ./memory/
memsearch index ./memory/ ./notes/ --provider google
memsearch index ./memory/ --force        # re-embed everything

Semantic Search โ€” search

Hybrid search (dense vector + BM25 full-text) with RRF reranking:

memsearch search "how to configure Redis caching"
memsearch search "auth flow" --top-k 10 --json-output

Live Sync โ€” watch

File watcher that auto-indexes on markdown changes (creates, edits, deletes):

memsearch watch ./memory/
memsearch watch ./memory/ ./notes/ --debounce-ms 3000

LLM Summarization โ€” compact

Compress indexed chunks into a condensed markdown summary using an LLM:

memsearch compact
memsearch compact --llm-provider anthropic --source ./memory/old-notes.md

Utilities โ€” stats / reset

memsearch stats                          # show total indexed chunk count
memsearch reset                          # drop all indexed data (with confirmation)

๐Ÿ“– Full command reference with all flags and examples โ†’ CLI Reference

๐Ÿ” How It Works

Markdown is the source of truth โ€” the vector store is just a derived index, rebuildable anytime.

  โ”Œโ”€โ”€โ”€ Search โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
  โ”‚                                                                    โ”‚
  โ”‚  "how to configure Redis?"                                         โ”‚
  โ”‚        โ”‚                                                           โ”‚
  โ”‚        โ–ผ                                                           โ”‚
  โ”‚   โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”     โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”     โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”   โ”‚
  โ”‚   โ”‚  Embed   โ”‚โ”€โ”€โ”€โ”€โ–ถโ”‚ Cosine similarityโ”‚โ”€โ”€โ”€โ”€โ–ถโ”‚ Top-K results    โ”‚   โ”‚
  โ”‚   โ”‚  query   โ”‚     โ”‚ (Milvus)        โ”‚     โ”‚ with source info โ”‚   โ”‚
  โ”‚   โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜     โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜     โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜   โ”‚
  โ”‚                                                                    โ”‚
  โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

  โ”Œโ”€โ”€โ”€ Ingest โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
  โ”‚                                                                    โ”‚
  โ”‚  MEMORY.md                                                         โ”‚
  โ”‚  memory/2026-02-09.md     โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”     โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”     โ”‚
  โ”‚  memory/2026-02-08.md โ”€โ”€โ”€โ–ถโ”‚ Chunker  โ”‚โ”€โ”€โ”€โ”€โ–ถโ”‚ Dedup          โ”‚     โ”‚
  โ”‚                           โ”‚(heading, โ”‚     โ”‚(chunk_hash PK) โ”‚     โ”‚
  โ”‚                           โ”‚paragraph)โ”‚     โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜     โ”‚
  โ”‚                           โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜             โ”‚              โ”‚
  โ”‚                                             new chunks only       โ”‚
  โ”‚                                                    โ–ผ              โ”‚
  โ”‚                                            โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”       โ”‚
  โ”‚                                            โ”‚  Embed &     โ”‚       โ”‚
  โ”‚                                            โ”‚  Milvus upsertโ”‚      โ”‚
  โ”‚                                            โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜       โ”‚
  โ”‚                                                                    โ”‚
  โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

  โ”Œโ”€โ”€โ”€ Watch โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
  โ”‚  File watcher (1500ms debounce) โ”€โ”€โ–ถ auto re-index / delete stale  โ”‚
  โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

  โ”Œโ”€โ”€โ”€ Compact โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
  โ”‚  Retrieve chunks โ”€โ”€โ–ถ LLM summarize โ”€โ”€โ–ถ write memory/YYYY-MM-DD.md โ”‚
  โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

๐Ÿ”’ The entire pipeline runs locally by default โ€” your data never leaves your machine unless you choose a remote backend or a cloud embedding provider.

๐Ÿงฉ Claude Code Plugin

memsearch ships with a Claude Code plugin โ€” a real-world example of agent memory in action. It gives Claude automatic persistent memory across sessions: every session is summarized to markdown, every prompt triggers a semantic search, and a background watcher keeps the index in sync. No commands to learn, no manual saving โ€” just install and go.

# 1. Set your embedding API key (OpenAI is the default provider)
export OPENAI_API_KEY="sk-..."

# 2. In Claude Code, add the marketplace and install the plugin
/plugin marketplace add zilliztech/memsearch
/plugin install memsearch

# 3. Restart Claude Code for the plugin to take effect, then start chatting!
claude

๐Ÿ“– Architecture, hook details, and development mode โ†’ Claude Code Plugin docs

โš™๏ธ Configuration

Settings are resolved in priority order (lowest โ†’ highest):

  1. Built-in defaults โ†’ 2. Global ~/.memsearch/config.toml โ†’ 3. Project .memsearch.toml โ†’ 4. CLI flags

API keys for embedding/LLM providers are read from standard environment variables (OPENAI_API_KEY, GOOGLE_API_KEY, VOYAGE_API_KEY, ANTHROPIC_API_KEY, etc.).

๐Ÿ“– Config wizard, TOML examples, and all settings โ†’ Getting Started โ€” Configuration

๐Ÿ”Œ Embedding Providers

Provider Install Default Model
OpenAI memsearch (included) text-embedding-3-small
Google memsearch[google] gemini-embedding-001
Voyage memsearch[voyage] voyage-3-lite
Ollama memsearch[ollama] nomic-embed-text
Local memsearch[local] all-MiniLM-L6-v2

๐Ÿ“– Provider setup and env vars โ†’ CLI Reference โ€” Embedding Provider Reference

๐Ÿ—„๏ธ Milvus Backend

memsearch supports three deployment modes โ€” just change milvus_uri:

Mode milvus_uri Best for
Milvus Lite (default) ~/.memsearch/milvus.db Personal use, dev โ€” zero config
Milvus Server http://localhost:19530 Multi-agent, team environments
Zilliz Cloud https://in03-xxx.api.gcp-us-west1.zillizcloud.com Production, fully managed

๐Ÿ“– Code examples and setup details โ†’ Getting Started โ€” Milvus Backends

๐Ÿ”— Integrations

memsearch works with any Python agent framework. Ready-made examples for:

  • LangChain โ€” use as a BaseRetriever in any LCEL chain
  • LangGraph โ€” wrap as a tool in a ReAct agent
  • LlamaIndex โ€” plug in as a custom retriever
  • CrewAI โ€” add as a tool for crew agents

๐Ÿ“– Copy-paste code for each framework โ†’ Integrations docs

๐Ÿ“š Links

  • Documentation โ€” full guides, API reference, and architecture details
  • Claude Code Plugin โ€” hook details, progressive disclosure, comparison with claude-mem
  • OpenClaw โ€” the memory architecture that inspired memsearch
  • Milvus โ€” the vector database powering memsearch
  • FAQ โ€” common questions and troubleshooting

Contributing

Bug reports, feature requests, and pull requests are welcome! See the Contributing Guide for development setup, testing, and plugin development instructions. For questions and discussions, join us on Discord.

๐Ÿ“„ License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

memsearch-0.1.12.tar.gz (2.8 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

memsearch-0.1.12-py3-none-any.whl (36.8 kB view details)

Uploaded Python 3

File details

Details for the file memsearch-0.1.12.tar.gz.

File metadata

  • Download URL: memsearch-0.1.12.tar.gz
  • Upload date:
  • Size: 2.8 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for memsearch-0.1.12.tar.gz
Algorithm Hash digest
SHA256 f434536d6819fa413c7c37ea583ed833cf86f54af78608a37a9ef5c929a7e413
MD5 dc2967eeab3eba95538b3d32ec54afc2
BLAKE2b-256 a2b81349d648f70b0cc46510a23736242897c0f0dc265ead84e911990f9ba307

See more details on using hashes here.

Provenance

The following attestation bundles were made for memsearch-0.1.12.tar.gz:

Publisher: release.yml on zilliztech/memsearch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file memsearch-0.1.12-py3-none-any.whl.

File metadata

  • Download URL: memsearch-0.1.12-py3-none-any.whl
  • Upload date:
  • Size: 36.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for memsearch-0.1.12-py3-none-any.whl
Algorithm Hash digest
SHA256 183c5a5ee1c3b4fcc42dd3b7e3db7280b6f9fbf33e03283fe03c733223f83cdd
MD5 5b3d8fa29f040d9e685460544f99897a
BLAKE2b-256 807ca8d914ed55c38196e92bed49fb4b99eeb9ed3c085d656d1953c6e05623f8

See more details on using hashes here.

Provenance

The following attestation bundles were made for memsearch-0.1.12-py3-none-any.whl:

Publisher: release.yml on zilliztech/memsearch

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page