Skip to main content

Universal memory layer for AI applications. Self-host in 5 minutes.

Project description

Remembra - AI Memory Layer

Persistent memory for AI applications. Self-host in 5 minutes.

What Is This?

Remembra is a universal memory layer for LLMs. It solves the fundamental problem that every AI forgets everything between sessions.

from remembra import Memory

memory = Memory(user_id="user_123")

# Store memories
memory.store("User prefers dark mode and works at Acme Corp")

# Recall with context
context = memory.recall("What are user's preferences?")
# Returns: "User prefers dark mode. Works at Acme Corp."

Why We're Building This

The Problem

Every AI app needs memory. Developers hack together solutions using vector databases, embeddings, and custom retrieval logic. It's complex, fragmented, and everyone rebuilds the same thing.

Current Solutions Suck

  • Mem0: $24M raised, but self-hosting docs are trash, pricing jumps from $19 to $249
  • Zep: Academic, complex to deploy
  • Letta: Not production-ready
  • LangChain Memory: Too basic, no persistence

Our Approach

  • Self-host in 5 minutes: One Docker command, everything bundled
  • Fair pricing: $0 → $29 → $99 (not $19 → $249)
  • Open source core: MIT license, own your data
  • Actually works: Built because we need it ourselves (Clawdbot)

Core Features

1. Simple Memory Operations

  • store() - Save memories with automatic extraction
  • recall() - Semantic search with context
  • update() - Intelligent merging
  • forget() - GDPR-compliant deletion

2. Entity Resolution (Our Killer Feature)

Knows that "Adam", "Adam Smith", "Mr. Smith", and "my husband" are the same person.

3. Temporal Awareness

Memories have time context. TTL support. Historical queries.

4. Hybrid Storage

Vector (semantic) + Graph (relationships) + Relational (metadata) in one system.

5. Observability Dashboard

See what's stored, debug retrievals, visualize entity graphs.

Quick Start

Self-Hosted (Recommended)

docker run -d -p 8787:8787 remembra/remembra

Python SDK

pip install remembra
from remembra import Memory

# Connect to local instance
memory = Memory(
    base_url="http://localhost:8787",
    user_id="user_123",
    project="my_app"
)

# Store
memory.store("User's name is John. He's a software engineer at Google.")

# Recall
context = memory.recall("Who is the user?")
print(context)
# "John is a software engineer at Google."

Documentation

Project Status

🚧 In Development - MVP target: 12 weeks

License

MIT License - Use it however you want.


Built by DolphyTech

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

remembra-0.2.0.tar.gz (26.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

remembra-0.2.0-py3-none-any.whl (33.3 kB view details)

Uploaded Python 3

File details

Details for the file remembra-0.2.0.tar.gz.

File metadata

  • Download URL: remembra-0.2.0.tar.gz
  • Upload date:
  • Size: 26.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.14

File hashes

Hashes for remembra-0.2.0.tar.gz
Algorithm Hash digest
SHA256 bf242d4e40b1ee67e459ae060de529925b1b37153004d441b3b58344c701e768
MD5 f09149cb83bc579e68d90e8673847f7f
BLAKE2b-256 091198a75e101568b230cb10dfe5249243f15b36bdb7ea77d7d809624d5a16a2

See more details on using hashes here.

File details

Details for the file remembra-0.2.0-py3-none-any.whl.

File metadata

  • Download URL: remembra-0.2.0-py3-none-any.whl
  • Upload date:
  • Size: 33.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.14

File hashes

Hashes for remembra-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 64ff39f49613eb0864be3c1aa6886c167a3122674ce6beda263695675de56c59
MD5 1197dc7e52621f3fc6e8ce465a277c01
BLAKE2b-256 92a572c9014ae78a14eae76cf668670b61285784ccd860e5664a4ce055a29de9

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page