Skip to main content

Moorcheh Universal Memory Layer for Agentic AI

Project description

MUMLA Logo

MUMLA - Moorcheh Universal Memory Layer for Agentic AI

Learn more · Join Discord

Moorcheh Discord License: MIT PyPI Version Twitter / X


What Is MUMLA?

MUMLA is a universal memory layer for agentic AI. While LLMs often forget context between sessions, MUMLA gives your agents long-term memory so they can carry context forward and remember what matters across sessions.

Why MUMLA Performs

MUMLA is built for teams that want high-quality agent memory without graph-heavy complexity. It combines immediate semantic availability, low-overhead serverless operation, and strong real-world memory accuracy so you can ship production workflows with a simpler architecture.

  • Zero-cost ingestion latency: No indexing wait or token usage at ingestion, so memories are available for retrieval immediately.
  • Zero storage cost at idle: Serverless architecture scales to zero when not in use.
  • State-of-the-art benchmark performance: Final evaluation results reached 89.8% on LongMemEval and 87.1% on LoCoMo.

🚀 MUMLA CLI

MUMLA comes with a powerful, developer-friendly Command Line Interface. You can initialize your environment, start the server, and manage your agent's memories completely from your terminal!

You need a Moorcheh API key to use MUMLA. Create one in the Moorcheh Dashboard.

MUMLA has native LLM access, so you don't need a separate external model API key for common memory workflows.

1. Install & Configure

pip install mumla

# Setup your environment (prompts for your Moorcheh API key)
mumla

2. Test Agent Memories

# Create and activate an agent session
mumla agent create customer-support
mumla agent activate customer-support

# Store memories with specific semantic types
mumla remember "The user prefers dark mode for the dashboard."
mumla remember "User's timezone is PST."

# Instantly recall relevant context
mumla recall "What mode does the user like?"

# Get grounded AI answers using built-in RAG
mumla answer "Based on the memory, what should the theme be set to?"

Supported Memory Types

instruction, fact, decision, goal, commitment, preference, relationship, context, event, learning, observation, artifact, error

Use memory types to categorize what you store so retrieval is cleaner and more controllable:

  • Save with a specific type: mumla remember "User prefers concise answers" --type preference
  • Filter by type when searching: mumla recall "user communication style" --type preference

Key Features

Capability Commands What it does
System status dashboard mumla status View environment, configuration, server health, active session, and registered agents.
Local server + web dashboard mumla serve, mumla ui Run the MUMLA API locally and open an interactive browser UI.
Agent lifecycle management mumla agent ... Create/list agents, activate/deactivate sessions, and run agent bootstrap for an intelligence snapshot.
Memory capture at scale mumla remember Store single memories with metadata or batch-ingest up to 100 records from JSON.
Advanced retrieval modes mumla recall Run standard search plus temporal queries (--as-of, --changed-since, --current-only) with filters.
Grounded QA over memory mumla answer Generate RAG answers using retrieved memory context.
Daily intelligence workflows mumla daily-summary, mumla conflicts Generate summaries, detect contradictions, and resolve conflicts interactively.
Session and automation controls mumla session ..., mumla schedule ... Inspect/extend sessions and enable scheduled daily summary runs.
Memory file pipelines mumla memory export, mumla memory sync Export structured memory markdown and sync MEMORY.md into projects.
Configuration inspection mumla config show Inspect API key status, active agent/session, server settings, and schedule time.
Multi-agent ecosystem integration mumla connect ... Connect/remove/list integrations for Claude Code, Codex, Cursor, Windsurf, Antigravity, Gemini CLI, Cline, Continue, OpenCode, Goose, Roo, GitHub Copilot, and Augment (local or global).

Additional setup guides are available at the Moorcheh YouTube channel.


🎯 REST API Endpoints

For programmatic access, MUMLA exposes a clean, session-based REST API.

Important: MUMLA does not have a hosted API server yet. To use these endpoints, run your own local server first:

cd mumla

# Start server
mumla serve

By default, call the endpoints on your local server (for example: "http://127.0.0.1:8000").

Agent Management

  • POST /api/v2/agents - Create a new agent namespace
  • GET /api/v2/agents - List all available agents
  • GET /api/v2/agents/{agent_id} - Get metadata for a specific agent
  • DELETE /api/v2/agents/{agent_id} - Delete an agent and all its memories

Session Management

  • POST /api/v2/agents/{agent_id}/activate - Start a session (returns a 6-hour JWT session_token)
  • POST /api/v2/agents/{agent_id}/deactivate - Manually end a session
  • GET /api/v2/session/current - Check the status/validity of the current session
  • POST /api/v2/session/extend - Extend the session expiration time

Memory Operations

  • POST /api/v2/agents/{agent_id}/remember - Store a new memory into the agent's semantic database
  • GET /api/v2/agents/{agent_id}/recall - Run an exact semantic search against the agent's memories
  • POST /api/v2/agents/{agent_id}/answer - Generate a grounded RAG answer based on the agent's memories

Authentication Required:

  • Authorization: Bearer {moorcheh_api_key} header
  • X-Session-Token: {session_token} header (for Session & Memory operations)

🤖 Why Moorcheh?

Moorcheh.ai - The world's only no-indexing semantic database.

The Revolutionary Difference

Traditional Vector DBs: Minutes of indexing delay, approximate search, stateful architecture

Moorcheh: Instant availability, exact search, serverless/stateless, 80% compute savings

Real Impact

Feature Traditional Moorcheh
Write-to-Search Minutes Instant
Accuracy Approximate Exact
Idle Costs Always running Zero
Free Tier Limited 100K ops/month

📞 Support & Documentation

Have questions or feedback? We're here to help:


MIT License

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

mumla-2.0.1.tar.gz (411.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

mumla-2.0.1-py3-none-any.whl (378.7 kB view details)

Uploaded Python 3

File details

Details for the file mumla-2.0.1.tar.gz.

File metadata

  • Download URL: mumla-2.0.1.tar.gz
  • Upload date:
  • Size: 411.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for mumla-2.0.1.tar.gz
Algorithm Hash digest
SHA256 526a8105da4f1d4af9edc7d54d430687d783a3efc1b5ea2b5116f78ff75099f9
MD5 bb5572401a58459ab6a8340e28730eb4
BLAKE2b-256 6ea4245df7fdca1f6c358ac1fb09836085fe11f7aba8995e0e29cf63d2b217f4

See more details on using hashes here.

File details

Details for the file mumla-2.0.1-py3-none-any.whl.

File metadata

  • Download URL: mumla-2.0.1-py3-none-any.whl
  • Upload date:
  • Size: 378.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for mumla-2.0.1-py3-none-any.whl
Algorithm Hash digest
SHA256 8c74da407389af85b898bea3b97813e25d2d3da8925b3ea9ab1ed9fa0be92525
MD5 2931afcb36abed1653a115e943524ab1
BLAKE2b-256 0a971b548aa0d66251a7f4396209ba0336893795e4e28efcb2ddd7fb3e22fa36

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page