Layered memory platform for AI agents โ log-first, RL-scored, consolidation-backed
Project description
๐ง extremis
Memory that gets smarter the more your agent uses it
One click ยท auto-provisions Postgres ยท memory persists across restarts
The problem
Every team building an AI agent hits the same wall.
Your agent forgets everything the moment a conversation ends. So you add memory. You set up a vector database, write chunking logic, figure out retrieval ranking, handle stale entries, add multi-user isolation. Three weeks later you've built a half-working RAG pipeline and still haven't shipped the actual feature.
And even when you ship it โ it doesn't learn. Every memory is treated identically. The fact your agent recalled a hundred times and the user loved sits next to one it got wrong once. Nothing improves. There's no feedback loop. You're running the same dumb cosine search forever.
The other problem is lock-in. Your vectors are in Pinecone. Moving them means re-embedding everything, rewriting your retrieval logic, and hoping nothing breaks.
extremis solves all three.
What makes extremis different
1. Memory that forgets intelligently
Every competitor focuses on storing memory. Nobody talks about forgetting.
Human memory doesn't keep everything forever โ unimportant things fade, important things strengthen. Agents with infinite, flat memory become slow and noisy over time. Intelligent forgetting is the hard problem nobody is solving.
extremis does two things here: recency decay (old memories rank lower automatically) and asymmetric RL weighting (negative feedback hurts 1.5ร more than positive feedback helps, because mistakes should leave a stronger mark). The result is a memory that naturally surfaces what matters and buries what doesn't.
mem = Extremis(config=Config(
recency_half_life_days=30, # episodic memories halve in rank every 30 days
rl_alpha=0.8, # strong RL signal โ useful things stick, useless things fade
))
# This memory will rank lower in every future search
mem.report_outcome([bad_memory_id], success=False, weight=1.0)
# โ score decreases by 1.5 (not 1.0 โ the asymmetry is intentional)
2. Memory that explains itself
Agents make decisions based on memory. But why did it recall that specific memory? Without explainability you're guessing, debugging is painful, and auditing is impossible.
Every recall() result includes a plain-English reason:
results = mem.recall("what does the user prefer?")
for r in results:
print(r.memory.content)
print(r.reason)
# "User prefers concise answers, no filler words"
# โ "similarity 0.91 ยท score +4.0 ยท used 8ร ยท 3d old"
# "User prefers dark mode in all UIs"
# โ "semantic (always included) ยท similarity 0.73 ยท score +1.0 ยท used 3ร ยท 12d old"
# "User once mentioned preferring email over Slack"
# โ "similarity 0.54 ยท score -1.5 ยท first recall ยท 45d old"
The reason tells you: how semantically relevant it was, how much feedback has validated it, how many times it's been used, and how old it is. Auditable. Debuggable.
3. Cross-agent shared memory
Right now memory is per-agent. But the next wave of AI is agent teams โ a research agent, a writing agent, a review agent, all working together. They need a shared brain.
extremis's namespace model already supports this. Multiple agents can read from and write to the same memory pool:
# All three agents share the same memory namespace
research = Extremis(config=Config(namespace="team_alpha"))
writer = Extremis(config=Config(namespace="team_alpha"))
reviewer = Extremis(config=Config(namespace="team_alpha"))
# Research agent stores what it found
research.remember("GPT-4 outperforms Claude on math benchmarks by 12%")
research.remember("Source: Stanford HAI report, April 2026")
# Writing agent recalls it without any extra wiring
results = writer.recall("GPT-4 performance data")
# โ [GPT-4 outperforms Claude on math benchmarks by 12%]
# โ [Source: Stanford HAI report, April 2026]
# Knowledge graph is shared too
research.kg_add_entity("Stanford HAI", EntityType.ORG)
research.kg_add_relationship("Stanford HAI", "HAI Report", "published")
print(writer.kg_query("Stanford HAI")) # same graph
4. No RAG pipeline to build
One pip install. Two lines of config. extremis handles embedding, storage, retrieval ranking, consolidation, and the knowledge graph. You call remember() and recall().
# Local โ zero infra
from extremis import Extremis
mem = Extremis()
# Your existing vector store
mem = Extremis(config=Config(store="pinecone", pinecone_api_key="..."))
# Self-hosted server โ no model download on the client
from extremis import HostedClient
mem = HostedClient(api_key="extremis_sk_...", base_url="http://your-server:8000")
# Same three lines work for all three
mem.remember("User is building a WhatsApp AI", conversation_id="c1")
results = mem.recall("what is the user building?")
mem.report_outcome([r.memory.id for r in results], success=True)
5. Backend portability โ no lock-in
Your vectors in Pinecone. Your team moves to Chroma. Your product needs Postgres. One command, everything migrates โ and re-embeds automatically if you're switching models:
extremis-migrate --from pinecone --to postgres \
--source-pinecone-api-key pk_... \
--dest-postgres-url postgresql://...
# Switching to OpenAI embeddings at the same time
extremis-migrate --from sqlite --to chroma \
--dest-embedder text-embedding-3-small
Coming soon
Memory health dashboard โ freshness score, contradiction count, retrieval hit rate, coverage gaps. Memory observability nobody is building yet.
Domain profiles โ pre-built memory configurations for common agent types:
# Coming in v0.2
from extremis.profiles import SalesAgent, CodingAgent, SupportAgent
mem = Extremis(profile=SalesAgent())
# Knows to remember: customer names, deal stage, objections, preferences
# Knows to forget: small talk after 7 days, meeting logistics after 24h
# Attention: high for "budget", "decision maker", "timeline"
How it works
The intelligence layer
extremis sits above your vector store. RL scoring, the knowledge graph, consolidation, and attention scoring are all backend-independent โ they work the same whether your vectors are in SQLite, Pinecone, or Chroma.
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ YOUR APP / AGENT โ
โ remember() ยท recall() ยท report_outcome() ยท kg_*() โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโฌโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ EXTREMIS INTELLIGENCE LAYER โ
โ RL scoring ยท Knowledge graph ยท Consolidation ยท Observer โ
โ Attention scorer ยท Namespace isolation ยท Log durability โ
โโโโโโฌโโโโโโโโโโโโโโโฌโโโโโโโโโโโโโโโฌโโโโโโโโโโโโโโโฌโโโโโโโโโโโโโโโ
โ โ โ โ
โโโโโโผโโโโ โโโโโโโโผโโโ โโโโโโโโผโโโ โโโโโโโโผโโโ
โSQLite โ โPostgres โ โ Chroma โ โPinecone โ
โ(local) โ โ+pgvectorโ โ (local) โ โ(hosted) โ
โโโโโโโโโโ โโโโโโโโโโโ โโโโโโโโโโโโ โโโโโโโโโโโ
The memory flow
Every conversation
โโโโโโโโโโโโโโโโโ
remember("user said X") โโโถ fsync to JSONL log (durable)
+ episodic memory (embedded + stored)
recall("topic") โโโถ embed query
โ identity + procedural (always included)
โ semantic + episodic (ranked by score)
โ ranked results
report_outcome(ids, +1/-1) โโโถ adjust utility scores
negative gets 1.5ร weight (human memory bias)
Periodically
โโโโโโโโโโโโ
consolidate() โโโถ read log since last checkpoint
โ Claude Haiku extracts facts
โ semantic/procedural memories written
โ checkpoint advanced (safe to re-run)
Retrieval ranking
Every recalled memory gets a final_rank that balances three signals:
final_rank = cosine_similarity
ร (1 + ฮฑ ยท tanh(utility_score)) โ learned from feedback
ร exp(โln2 ยท age_days / half_life) โ recency decay
A memory that has proven useful (+1 feedback) ranks above an equally similar but unvalidated memory. Negative signals apply 1.5ร weight โ the same asymmetry human threat-learning uses.
Memory layers
| Layer | What it holds | Written by | Always recalled? |
|---|---|---|---|
identity |
Who the user fundamentally is | Human review only | โ Always |
procedural |
Behavioural rules: "ask about deadline first" | Consolidator | โ Always |
semantic |
Durable facts: "user is a solo Python developer" | Consolidator | By relevance |
episodic |
Timestamped conversation events | remember() |
By relevance |
working |
Session-scoped, expires on a set datetime | remember_now() |
By relevance |
Knowledge graph
Beyond vectors, extremis maintains a structured graph โ answers structural questions that semantic search can't:
mem.kg_add_entity("Alice", EntityType.PERSON)
mem.kg_add_entity("Acme Corp", EntityType.ORG)
mem.kg_add_relationship("Alice", "Acme Corp", "works_at", weight=0.95)
mem.kg_add_attribute("Alice", "timezone", "Asia/Dubai")
mem.kg_add_attribute("Alice", "tone", "formal")
# "Who does Alice work for?" โ can't answer with cosine similarity alone
result = mem.kg_query("Alice")
# โ Entity + all relationships + all attributes + BFS traverse
# Two-hop traverse
graph = mem.kg_traverse("Alice", depth=2)
Attention scoring
Before deciding how much to engage with an incoming message, score it โ free, zero LLM cost:
score = sender_score + channel_score + content_score + context_score (0โ100)
full โฅ 75 โ engage fully
standard โฅ 50 โ balanced response
minimal โฅ 25 โ brief acknowledgement
ignore < 25 โ skip
Observer (log compression)
Compresses raw log entries into priority-tagged observations โ no LLM, runs instantly:
๐ด CRITICAL decisions, errors, deadlines, shipped/launched, reward signals
๐ก CONTEXT reasons, insights, learnings, "because", "discovered"
๐ข INFO everything else
Install
Requires Python 3.11+
If
pip install extremissays "no matching distribution found" โ your defaultpippoints to Python 3.9 or older. This is common on macOS.Check your version:
python3 --version
Platform Fix macOS brew install python@3.11then usepip3.11Linux sudo apt install python3.11 python3.11-pipWindows python.org/downloads
# Confirm you have Python 3.11+
python3.11 --version
# Core โ SQLite + local sentence-transformers (no API key needed)
pip3.11 install extremis
# + MCP server (Claude Desktop / Code)
pip3.11 install "extremis[mcp]"
# + Postgres backend
pip3.11 install "extremis[postgres]"
# + Chroma backend
pip3.11 install "extremis[chroma]"
# + Pinecone backend
pip3.11 install "extremis[pinecone]"
# + OpenAI embeddings (swap out the 90 MB model download)
pip3.11 install "extremis[openai]"
# + Hosted API server
pip3.11 install "extremis[server]"
# + Python SDK for hosted cloud
pip3.11 install "extremis[client]"
# Everything
pip3.11 install "extremis[all]"
Requires Python 3.11+
First run note โ
sentence-transformersdownloadsall-MiniLM-L6-v2(~90 MB) on first use. One-time, cached to~/.cache/huggingface/. To skip it, use OpenAI embeddings:EXTREMIS_EMBEDDER=text-embedding-3-small.
Quick start
from extremis import Extremis, MemoryLayer
from extremis.types import EntityType
mem = Extremis() # ~/.extremis/ by default
# โโ Remember โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
mem.remember("User is building a WhatsApp AI", conversation_id="conv_001")
mem.remember("User prefers concise answers", conversation_id="conv_001")
# Skip the log for time-sensitive or high-confidence facts
mem.remember_now(
"Flight departs Thursday at 06:00",
layer=MemoryLayer.EPISODIC,
confidence=0.99,
)
# โโ Recall โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
results = mem.recall("what product is the user building?", limit=5)
for r in results:
print(f"[{r.memory.layer.value}] {r.memory.content} rank={r.final_rank:.3f}")
# โโ Feedback โ memories get smarter over time โโโโโโโโโโโโโโโโโ
mem.report_outcome([r.memory.id for r in results[:2]], success=True)
# โโ Knowledge graph โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
mem.kg_add_entity("User", EntityType.PERSON)
mem.kg_add_entity("Friday", EntityType.PROJECT)
mem.kg_add_relationship("User", "Friday", "building")
mem.kg_add_attribute("User", "timezone", "Asia/Dubai")
print(mem.kg_query("User"))
# โโ Attention scoring โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
result = mem.score_attention("URGENT: the API is down!", channel="dm")
print(result.level) # โ "full"
print(result.score) # โ 85
# โโ Consolidation (nightly / on-demand) โโโโโโโโโโโโโโโโโโโโโโโ
from extremis.consolidation import LLMConsolidator
consolidator = LLMConsolidator(mem._config, mem._embedder)
r = consolidator.run_pass(mem.get_log(), mem.get_local_store(), mem.get_local_store())
print(f"{r.memories_created} facts extracted from logs")
Storage backends
All backends share the same API. Swap with one env var.
Don't want anything stored locally?
Three options โ all work out of the box:
| Option | Local footprint | Cost |
|---|---|---|
| Postgres on Supabase / Neon | None | Free tier available |
| Pinecone | RL score sidecar only (~KB) | Free tier available |
| HostedClient (your own server) | None at all | Your hosting cost |
Quickest: free Postgres on Supabase
# 1. Create project at supabase.com, grab the connection string
# 2. Enable pgvector: run "CREATE EXTENSION vector;" in the SQL editor
pip3.11 install "extremis[postgres]"
EXTREMIS_STORE=postgres EXTREMIS_POSTGRES_URL=postgresql://... python3.11 your_app.py
Zero footprint: HostedClient
from extremis import HostedClient
# deploy extremis-server on Railway/Fly/Render, point at it
mem = HostedClient(api_key="extremis_sk_...", base_url="https://your-server.railway.app")
# nothing written locally โ not even the embedding model
SQLite โ default, zero infrastructure
EXTREMIS_STORE=sqlite
EXTREMIS_FRIDAY_HOME=~/.extremis # DB at ~/.extremis/local.db
Postgres + pgvector โ production scale, ranking in SQL
pip3.11 install "extremis[postgres]"
EXTREMIS_STORE=postgres
EXTREMIS_POSTGRES_URL=postgresql://user:pass@host/extremis
Requires CREATE EXTENSION vector; in your database. Schema migrates automatically on first start.
Chroma โ local vector DB, great for teams
pip3.11 install "extremis[chroma]"
EXTREMIS_STORE=chroma
EXTREMIS_CHROMA_PATH=~/.extremis/chroma
Pinecone โ serverless hosted vectors
pip3.11 install "extremis[pinecone]"
EXTREMIS_STORE=pinecone
EXTREMIS_PINECONE_API_KEY=pk_...
EXTREMIS_PINECONE_INDEX=extremis
Create the index first (dimension must match your embedder):
from pinecone import Pinecone, ServerlessSpec
pc = Pinecone(api_key="pk_...")
pc.create_index("extremis", dimension=384, metric="cosine",
spec=ServerlessSpec(cloud="aws", region="us-east-1"))
OpenAI embeddings โ no model download
pip3.11 install "extremis[openai]"
EXTREMIS_EMBEDDER=text-embedding-3-small
OPENAI_API_KEY=sk-...
EXTREMIS_EMBEDDING_DIM=1536
Works with any storage backend. Removes the 90 MB local model download.
Migrating backends
Move all memories between backends in one command. extremis re-embeds automatically if the source and destination use different embedding models.
pip3.11 install "extremis[chroma,pinecone]"
# Escape Pinecone lock-in โ local SQLite
extremis-migrate --from pinecone --to sqlite \
--source-pinecone-api-key pk_... \
--source-pinecone-index my-index
# Local SQLite โ Postgres (upgrade to production)
extremis-migrate --from sqlite --to postgres \
--dest-postgres-url postgresql://...
# Switch to OpenAI embeddings while migrating
extremis-migrate --from sqlite --to chroma \
--dest-embedder text-embedding-3-small
# Dry run โ count what would be migrated
extremis-migrate --from sqlite --to chroma --dry-run
Hosted API
Run extremis as a service โ your users call it with an API key, all compute happens server-side. No model download on the client. No local database.
Status: The server is fully built and self-hostable today. A managed cloud at
api.extremis.comis in progress โ join the waitlist.
One-click deploy to Render (memory lives in Render Postgres)
Clicking this button deploys extremis-server and provisions a free Postgres database automatically via render.yaml. Memory lives in Render's managed Postgres โ persistent across restarts and redeploys.
Getting your API key โ check the logs, it's already there.
On first startup, extremis auto-generates a key and prints it in the server logs. In Render:
- Click your extremis service โ Logs tab
- Look for the block that says
extremis โ FIRST START - Copy the key that starts with
extremis_sk_...
============================================================
extremis โ FIRST START
============================================================
No API keys found. Generated your first key:
extremis_sk_xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
Namespace: default
Store this key โ it will NOT be shown again.
============================================================
Connect from anywhere with zero local footprint:
from extremis import HostedClient
mem = HostedClient(api_key="extremis_sk_...", base_url="https://your-app.onrender.com")
To create additional keys (e.g. per user/namespace), use Render's Shell tab:
extremis-server create-key --namespace alice --label "alice prod"
Deploy to Railway (manual โ 3 steps)
โ ๏ธ Don't use SQLite on Railway. Container filesystems are ephemeral โ memories are lost on every restart. Always use Railway Postgres.
- Create a new project on railway.app โ Deploy from GitHub repo โ select
extremis - Add a Postgres plugin: + New โ Database โ PostgreSQL
- Set these environment variables on the extremis service:
EXTREMIS_STORE=postgres EXTREMIS_POSTGRES_URL=${{Postgres.DATABASE_URL}}
Railway injects the URL automatically. Memory now lives in Railway's managed Postgres.
Self-host locally in 2 minutes
pip3.11 install "extremis[server]"
# Generate an API key
extremis-server create-key --namespace alice --label "prod"
# โ extremis_sk_xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx (shown once, store it)
# Start the server
extremis-server serve --host 0.0.0.0 --port 8000
# Or with Docker (bundles Postgres + pgvector)
docker compose up
Connect from Python
from extremis import HostedClient
# Point at your self-hosted server
mem = HostedClient(api_key="extremis_sk_...", base_url="http://your-server:8000")
# Exact same API as Memory โ nothing else changes
mem.remember("User is building a WhatsApp AI", conversation_id="c1")
results = mem.recall("WhatsApp")
mem.report_outcome([r.memory.id for r in results], success=True)
API endpoints
POST /v1/memories/remember append to log + episodic store
POST /v1/memories/recall semantic search, layered retrieval
POST /v1/memories/report RL signal (+1/โ1)
POST /v1/memories/store direct write to any layer
POST /v1/memories/consolidate LLM consolidation pass
GET /v1/memories/observe priority-tagged log compression
POST /v1/kg/write add entity / relationship / attribute
POST /v1/kg/query query + BFS graph traverse
POST /v1/attention/score 0โ100 message priority score
GET /v1/health
All requests require Authorization: Bearer extremis_sk_.... Namespace is derived from the key.
Key management
extremis-server create-key --namespace prod_user_123 --label "production"
extremis-server list-keys
extremis-server list-keys --namespace prod_user_123
extremis-server revoke-key --key-hash abc123...
Deploy to production
Railway / Render (fastest โ 10 minutes):
- Point at the
Dockerfile - Set
EXTREMIS_STORE=postgresandEXTREMIS_POSTGRES_URL - Deploy
Fly.io:
fly launch
fly secrets set EXTREMIS_STORE=postgres EXTREMIS_POSTGRES_URL=postgresql://...
fly deploy
Self-hosted Docker:
docker build -t extremis-server .
docker run -p 8000:8000 \
-e EXTREMIS_STORE=postgres \
-e EXTREMIS_POSTGRES_URL=postgresql://... \
-v lore_data:/data \
extremis-server
MCP setup
Claude Desktop
pip3.11 install "extremis[mcp]"
Add to ~/Library/Application Support/Claude/claude_desktop_config.json:
{
"mcpServers": {
"extremis": {
"command": "extremis-mcp",
"env": {
"EXTREMIS_FRIDAY_HOME": "~/.extremis",
"ANTHROPIC_API_KEY": "sk-ant-..."
}
}
}
}
Restart Claude Desktop. Nine tools appear automatically.
Claude Code
claude mcp add extremis extremis-mcp \
--env EXTREMIS_FRIDAY_HOME=~/.extremis \
--env ANTHROPIC_API_KEY=sk-ant-...
SSE / HTTP mode
extremis-mcp --transport sse --port 8765
MCP tools
| Tool | What it does | LLM cost |
|---|---|---|
memory_remember |
Append to log + episodic store | None |
memory_recall |
Semantic search, identity+procedural always included | None |
memory_report_outcome |
+1/โ1 RL signal on recalled memories | None |
memory_remember_now |
Direct write to any layer (bypass log) | None |
memory_consolidate |
Distil logs into semantic/procedural memories | Haiku |
memory_kg_write |
Add entity / relationship / attribute | None |
memory_kg_query |
Query entity + BFS graph traverse | None |
memory_observe |
Compress log into ๐ด๐ก๐ข observations | None |
memory_score_attention |
Score a message 0โ100 | None |
Multi-user / namespace isolation
Two isolation models:
Instance-level โ each user gets their own process and EXTREMIS_FRIDAY_HOME. What Claude Desktop does naturally.
Namespace-level โ one deployment, many users. All memories, logs, and graph data scoped per namespace. Zero leakage.
EXTREMIS_NAMESPACE=alice extremis-mcp # Alice's memory
EXTREMIS_NAMESPACE=bob extremis-mcp # Bob's โ completely separate, same DB
mem_alice = Extremis(config=Config(namespace="alice"))
mem_bob = Extremis(config=Config(namespace="bob"))
# same DB file, zero crossover
Configuration
All settings via EXTREMIS_ environment variables or a .env file:
| Variable | Default | Description |
|---|---|---|
EXTREMIS_STORE |
sqlite |
Backend: sqlite ยท postgres ยท chroma ยท pinecone |
EXTREMIS_NAMESPACE |
default |
User/agent isolation scope |
EXTREMIS_FRIDAY_HOME |
~/.extremis |
Base dir for logs and SQLite DB |
EXTREMIS_POSTGRES_URL |
(empty) | Postgres DSN (required when store=postgres) |
EXTREMIS_CHROMA_PATH |
~/.extremis/chroma |
ChromaDB persistence directory |
EXTREMIS_PINECONE_API_KEY |
(empty) | Pinecone API key |
EXTREMIS_PINECONE_INDEX |
extremis |
Pinecone index name |
EXTREMIS_EMBEDDER |
all-MiniLM-L6-v2 |
Model name โ sentence-transformers or OpenAI |
EXTREMIS_EMBEDDING_DIM |
384 |
Vector dimension (must match model) |
EXTREMIS_OPENAI_API_KEY |
(empty) | OpenAI key (required for OpenAI embedders) |
EXTREMIS_CONSOLIDATION_MODEL |
claude-haiku-4-5-20251001 |
LLM for consolidation |
EXTREMIS_RL_ALPHA |
0.5 |
Utility score weight in retrieval ranking |
EXTREMIS_RECENCY_HALF_LIFE_DAYS |
90 |
Recency decay half-life |
EXTREMIS_ATTENTION_FULL_THRESHOLD |
75 |
Score โฅ this โ full attention |
EXTREMIS_ATTENTION_STANDARD_THRESHOLD |
50 |
Score โฅ this โ standard |
EXTREMIS_ATTENTION_MINIMAL_THRESHOLD |
25 |
Score โฅ this โ minimal |
How it compares
| extremis | Mem0 | LangChain | Zep | Raw Pinecone | |
|---|---|---|---|---|---|
| Self-hostable | โ | โ cloud only | โ | โ | โ |
| Backend-agnostic | โ 4 backends | โ | โ ๏ธ manual | โ | โ |
| RL-scored retrieval | โ | โ | โ | โ | โ |
| Asymmetric feedback (1.5ร) | โ | โ | โ | โ | โ |
| Knowledge graph | โ | โ | โ | โ | โ |
| 5-layer memory | โ | โ ๏ธ basic | โ ๏ธ basic | โ ๏ธ basic | โ |
| Log-first durability | โ | โ | โ | โ | โ |
| Migration CLI | โ | โ | โ | โ | โ |
| Attention scoring | โ | โ | โ | โ | โ |
| MCP server (Claude) | โ | โ | โ | โ | โ |
| Hosted API | โ self-host | โ | โ | โ | โ |
| Open source | โ MIT | โ ๏ธ partial | โ | โ | โ |
Project structure
extremis/
โโโ src/extremis/
โ โโโ api.py โ Memory โ the local API
โ โโโ client.py โ HostedClient โ the cloud API (same interface)
โ โโโ config.py โ Config (EXTREMIS_ env vars)
โ โโโ types.py โ Memory, Entity, Observation, AttentionResult, ...
โ โโโ interfaces.py โ LogStore, MemoryStore, Embedder protocols
โ โโโ migrate.py โ Migrator + extremis-migrate CLI
โ โโโ storage/
โ โ โโโ sqlite.py โ SQLiteMemoryStore
โ โ โโโ postgres.py โ PostgresMemoryStore (pgvector, ranking in SQL)
โ โ โโโ chroma.py โ ChromaMemoryStore
โ โ โโโ pinecone_store.py โ PineconeMemoryStore
โ โ โโโ kg.py โ SQLiteKGStore
โ โ โโโ log.py โ FileLogStore (JSONL, fsync, checkpoints)
โ โ โโโ score_index.py โ SQLiteScoreIndex (RL scores for external backends)
โ โโโ embeddings/
โ โ โโโ sentence_transformers.py
โ โ โโโ openai.py
โ โโโ consolidation/
โ โ โโโ consolidator.py โ LLMConsolidator (log โ Claude Haiku โ memories)
โ โ โโโ prompts.py
โ โโโ observer/
โ โ โโโ observer.py โ HeuristicObserver (๐ด๐ก๐ข)
โ โโโ scorer/
โ โ โโโ attention.py โ AttentionScorer (0โ100)
โ โโโ mcp/
โ โ โโโ server.py โ FastMCP server (9 tools)
โ โโโ server/
โ โโโ app.py โ FastAPI hosted API
โ โโโ auth.py โ API key management
โ โโโ deps.py โ FastAPI dependencies
โ โโโ routes/ โ memories, kg, health
โโโ Dockerfile
โโโ docker-compose.yml
โโโ tests/ โ 50 test files, no LLM calls
Contributing
See CONTRIBUTING.md. The quickest contribution is a new storage backend โ implement the MemoryStore protocol in storage/ and add tests. We'll merge it.
Security
See SECURITY.md for reporting vulnerabilities.
License
MIT ยท Built by Ashwani Jha
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file extremis-0.2.4.tar.gz.
File metadata
- Download URL: extremis-0.2.4.tar.gz
- Upload date:
- Size: 94.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
7c8e48cfaf5b9e08fc8851ca83971fe31afb0fa1eb483f52a102bdd13987cdd8
|
|
| MD5 |
8bfaa271f3dc5f919f18147122b650c8
|
|
| BLAKE2b-256 |
a0a8d9e0f1d1eecf46477d31f769dfccbf60faeba1a6d3e18521ad9683357a9a
|
Provenance
The following attestation bundles were made for extremis-0.2.4.tar.gz:
Publisher:
publish.yml on ashwanijha04/extremis
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
extremis-0.2.4.tar.gz -
Subject digest:
7c8e48cfaf5b9e08fc8851ca83971fe31afb0fa1eb483f52a102bdd13987cdd8 - Sigstore transparency entry: 1469347803
- Sigstore integration time:
-
Permalink:
ashwanijha04/extremis@612c1d1dce4f92f7c3e4cdecbce0a353f9f11f30 -
Branch / Tag:
refs/tags/v0.2.4 - Owner: https://github.com/ashwanijha04
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@612c1d1dce4f92f7c3e4cdecbce0a353f9f11f30 -
Trigger Event:
push
-
Statement type:
File details
Details for the file extremis-0.2.4-py3-none-any.whl.
File metadata
- Download URL: extremis-0.2.4-py3-none-any.whl
- Upload date:
- Size: 76.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
79f36b2d90beb9727e25b01ba2940e459de75998f76dfe841efa8485a23d8de6
|
|
| MD5 |
e0893ca20b8490e5d130ad48f4ad47cb
|
|
| BLAKE2b-256 |
47fe5c4e0d146b216c838e36c439979e2eaafb8cb4980d7ec4778b8432248ec3
|
Provenance
The following attestation bundles were made for extremis-0.2.4-py3-none-any.whl:
Publisher:
publish.yml on ashwanijha04/extremis
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
extremis-0.2.4-py3-none-any.whl -
Subject digest:
79f36b2d90beb9727e25b01ba2940e459de75998f76dfe841efa8485a23d8de6 - Sigstore transparency entry: 1469347898
- Sigstore integration time:
-
Permalink:
ashwanijha04/extremis@612c1d1dce4f92f7c3e4cdecbce0a353f9f11f30 -
Branch / Tag:
refs/tags/v0.2.4 - Owner: https://github.com/ashwanijha04
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@612c1d1dce4f92f7c3e4cdecbce0a353f9f11f30 -
Trigger Event:
push
-
Statement type: