Skip to main content

Universal memory layer for AI applications. Self-host in 5 minutes.

Project description

Remembra - AI Memory Layer

Persistent memory for AI applications. Self-host in 5 minutes.

What Is This?

Remembra is a universal memory layer for LLMs. It solves the fundamental problem that every AI forgets everything between sessions.

from remembra import Memory

memory = Memory(user_id="user_123")

# Store memories
memory.store("User prefers dark mode and works at Acme Corp")

# Recall with context
context = memory.recall("What are user's preferences?")
# Returns: "User prefers dark mode. Works at Acme Corp."

Why We're Building This

The Problem

Every AI app needs memory. Developers hack together solutions using vector databases, embeddings, and custom retrieval logic. It's complex, fragmented, and everyone rebuilds the same thing.

Current Solutions Suck

  • Mem0: $24M raised, but self-hosting docs are trash, pricing jumps from $19 to $249
  • Zep: Academic, complex to deploy
  • Letta: Not production-ready
  • LangChain Memory: Too basic, no persistence

Our Approach

  • Self-host in 5 minutes: One Docker command, everything bundled
  • Fair pricing: $0 → $29 → $99 (not $19 → $249)
  • Open source core: MIT license, own your data
  • Actually works: Built because we need it ourselves (Clawdbot)

Core Features

1. Simple Memory Operations

  • store() - Save memories with automatic extraction
  • recall() - Semantic search with context
  • update() - Intelligent merging
  • forget() - GDPR-compliant deletion

2. Entity Resolution (Our Killer Feature)

Knows that "Adam", "Adam Smith", "Mr. Smith", and "my husband" are the same person.

3. Temporal Awareness

Memories have time context. TTL support. Historical queries.

4. Hybrid Storage

Vector (semantic) + Graph (relationships) + Relational (metadata) in one system.

5. Observability Dashboard

See what's stored, debug retrievals, visualize entity graphs.

Quick Start

Self-Hosted (Recommended)

docker run -d -p 8787:8787 remembra/remembra

Python SDK

pip install remembra
from remembra import Memory

# Connect to local instance
memory = Memory(
    base_url="http://localhost:8787",
    user_id="user_123",
    project="my_app"
)

# Store
memory.store("User's name is John. He's a software engineer at Google.")

# Recall
context = memory.recall("Who is the user?")
print(context)
# "John is a software engineer at Google."

Documentation

Project Status

🚧 In Development - MVP target: 12 weeks

License

MIT License - Use it however you want.


Built by DolphyTech

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

remembra-0.6.2.tar.gz (89.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

remembra-0.6.2-py3-none-any.whl (98.6 kB view details)

Uploaded Python 3

File details

Details for the file remembra-0.6.2.tar.gz.

File metadata

  • Download URL: remembra-0.6.2.tar.gz
  • Upload date:
  • Size: 89.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.14

File hashes

Hashes for remembra-0.6.2.tar.gz
Algorithm Hash digest
SHA256 29145e5454a768a9ab017af12fe26afdf75a303d9002bf979e0c4bb9a377309d
MD5 3e0058eb7785352a57af5bd28d1f9ed1
BLAKE2b-256 7b4d773fc0857b4b7559da1ba09a70f0edf87eb3f30fbd80f8fbda95b7201567

See more details on using hashes here.

File details

Details for the file remembra-0.6.2-py3-none-any.whl.

File metadata

  • Download URL: remembra-0.6.2-py3-none-any.whl
  • Upload date:
  • Size: 98.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.14

File hashes

Hashes for remembra-0.6.2-py3-none-any.whl
Algorithm Hash digest
SHA256 424706df7a7bb8130a1db3163e83eb0cbb111f39acc7401d30bff8299464a38b
MD5 92c589e9b798e60a045fda28fb8b1c10
BLAKE2b-256 b14c725c74a3f689876ba4d6fb28d99f14deee1b188447effaea6f5e12f8dce9

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page