Brain-inspired AI memory compression with optional post-quantum encryption. Extends context windows 3-50x through tiered storage and semantic indexing.
Project description
Engram
Some parts of Engram are working, others are still in testing. Tiered compression, keyword search, and the Merkle integrity layer are stable and in daily use. Post-quantum encryption (ML-KEM-768) works but hasn't been independently audited. CoGraph (associative recall via co-occurrence + spreading activation) is new and being evaluated. Working memory encryption is not yet solved — the retrieval layer needs to reliably predict what the LLM needs before selective decryption is viable. This is a research project. The code is open because the ideas might be useful to others building AI memory infrastructure. Use at your own discretion.
Brain-modeled tiered memory for AI — with post-quantum encryption and Merkle integrity.
The Problem
AI assistants forget. Context windows have hard limits. Fill it up and the oldest context falls off. Your assistant loses decisions from last week, code reviews from last month, and architecture conversations from last quarter. You re-explain the same context every session.
That's today. Now think about a decade from now. Persistent AI memory is coming — assistants that remember years of your work, your preferences, your decisions. That's terabytes of session data accumulating on your machine. Most users don't have the storage for it. And the ones who do have all of it sitting in plaintext — every conversation, every code review, every personal decision — uncompressed, unencrypted, unmanageable.
The standard solution is "just expand the context window." But larger windows don't solve finding the right context (10M tokens of everything is worse than 32K of the right thing), don't solve cost (loading gigabytes per API call is expensive), don't solve storage (years of sessions will consume hundreds of gigabytes uncompressed), and don't solve security (more data per prompt = more data at risk).
The Approach
This started as a hardware constraint problem. Limited disk, limited context window, months of AI session data piling up. The question was: how do you keep months of AI memory accessible without unlimited storage?
The answer came from neuroscience.
Your brain doesn't hold everything in working memory. It uses a tiered architecture: the prefrontal cortex holds ~7 items for immediate access. The hippocampus consolidates recent experiences into longer-term storage. The neocortex stores long-term memories distributed across regions, reconstructed from fragments when the right cue triggers recall. And the deepest memories take real effort to surface — the tip-of-tongue phenomenon.
Each tier trades retrieval speed for storage efficiency. That's not a limitation. It's what lets the system scale across decades. Engram applies this same architecture to AI memory.
HOT (now) ████████████████████████████████████████ 1,500 KB 1x instant
WARM (1w) ████████████ 340 KB 4-5x ~10ms
COLD (1mo) ██████ 150 KB 8-12x ~200ms
FROZEN (3mo) ██ 50 KB 20-50x ~2s
Three Problems Solved
Context persistence. Old sessions compress into searchable tiers instead of being dropped. Months of decisions, code reviews, and conversations stay accessible through the semantic index. No re-explaining.
Hardware constraints. Six months of daily sessions would consume gigabytes uncompressed. Engram's multi-stage pipeline (4-5x warm, 8-12x cold, 20-50x frozen) shrinks that footprint by 90%+ while keeping everything searchable. Smarter storage, not more storage.
Security of AI memory. This is the problem nobody is solving yet. Every session you've ever had with an AI — stored in plaintext on disk. As personal AI assistants become persistent (remembering your preferences, your codebase, your decisions), that memory becomes a high-value target. An adversary who accesses your AI memory gets your entire professional history.
Engram adds post-quantum encryption (ML-KEM-768 + X25519 hybrid, NIST FIPS 203) so your AI memory is protected against both current and future threats. The "harvest now, decrypt later" (HNDL) risk — where an adversary records encrypted data today and breaks it with a quantum computer in 10-20 years — is real for long-lived data. AI memory that persists for years needs encryption that persists for decades. Per-artifact keys. Per-tier keypairs. Private keys handled by a compiled Rust sidecar that never lets them enter Python's address space.
Every recalled artifact is verified via a SHA3-256 Merkle tree with HMAC root sealing — proving the memory hasn't been tampered with and that the AI isn't hallucinating past conversations. Integrity and confidentiality, not just one or the other.
Context drift — the compounding approximation error that occurs when an LLM references compressed or summarized memory instead of original content. Each recall through a lossy representation introduces small errors. Over many references, these errors compound: memory A gets summarized, memory B references A's summary, memory C references B which referenced A's summary. C is now two approximation steps from the original. The model builds confidently on a drifting foundation.
Engram's Merkle tree addresses the integrity dimension of context drift: every recalled artifact is provably identical to what was originally stored. The proof says "this is the real Session 47, bit-for-bit." The summary may lose nuance (semantic drift from compression is inherent), but the underlying artifact is provably intact. When the summary isn't enough, the original content is one engram recall away — decompressed, decrypted, verified, and promoted back to hot tier.
AI is here to stay. Secure memory persistence at scale — preventing context drift, protecting against HNDL, and enabling LLMs to reference back to verified ground truth — is the next big infrastructure problem. Engram is a first step.
No additional setup needed to expand memory. Run engram init (auto-detects your AI assistants), then engram run (compresses and indexes everything). Your AI immediately has access to months of context through the semantic index. The context window didn't get bigger. The memory behind it got smarter.
HOT (now) ████████████████████████████████████████ 1,500 KB 1x instant
WARM (1w) ████████████ 340 KB 4-5x ~10ms
COLD (1mo) ██████ 150 KB 8-12x ~500ms
FROZEN (3mo) ██ 50 KB 20-50x ~5s
Merkle Tree Integrity — Anti-Hallucination for AI Memory
Added 2026-03-21. SHA-256 replaced entirely with SHA3-256. HMAC-SHA3 root sealing added. Python implementation archived — all Merkle operations moved to Rust sidecar.
Two separate protections: SHA3-256 Merkle tree proves data hasn't been tampered with (integrity). HMAC seal proves who computed it (authentication). Integrity without authentication means an attacker can build a valid tree with fake data. Authentication without integrity means the key holder could tamper with artifacts. You need both. Neither replaces Engram's existing encryption (ML-KEM-768 + AES-256-GCM) — encryption hides content, Merkle proves content is real. Separate threats, separate systems.
AI systems hallucinate. They fabricate citations, invent past conversations, and confidently reference decisions that never happened. When your AI says "we decided X last week," how do you know that conversation actually existed?
Engram's Merkle tree proves it.
Every artifact registered in Engram gets a leaf in a binary hash tree. The root hash is a single value that covers every artifact across all four tiers. If any artifact is tampered with, corrupted, or fabricated, the root hash changes.
Root: 061e... (SHA3-256)
Seal: 145f... (HMAC-SHA3, PQC-derived key)
/ \
Hash AB Hash CD
/ \ / \
Hash A Hash B Hash C Hash D
| | | |
Session 1 Session 2 Session 3 Session 4
(hot) (warm) (cold) (frozen)
Anti-hallucination: When the AI claims "we discussed this in Session 3," Engram can produce a Merkle proof — a path of hashes from Session 3's leaf to the root. If the proof verifies, the conversation is real. If it doesn't, the AI is hallucinating. Cryptographic proof, not trust.
Selective disclosure: To prove Session 3 exists, you only need 3 hashes (the proof path), not the contents of Sessions 1, 2, or 4. You prove the memory is real without revealing anything else.
Tamper detection: Run engram verify. One root hash check tells you every artifact across all four tiers is intact. No need to decompress cold or frozen files. O(log n) instead of O(n).
$ engram verify
Integrity Verification
========================================
Total checked: 487
Passed: 482
Failed: 0
Skipped: 5 (no hash stored)
Merkle Tree
Root: 061e4093b1a0...
Leaves: 482
Integrity: PASS
How it works with the tiers:
| Tier | Merkle Behavior |
|---|---|
| Hot | Leaf added on engram scan. Hash = SHA3-256 of original content. |
| Warm | Leaf preserved. Hash stays valid — compression is lossless at this tier. |
| Cold | Leaf preserved. Hash of pre-compression content remains verifiable via proof. |
| Frozen | Leaf preserved. Months-old artifacts verifiable without decompressing Parquet. |
Why SHA3-256 Instead of SHA-256
Most Merkle tree implementations use SHA-256. Bitcoin established it as the default in 2009, and RFC 6962 (Certificate Transparency) codified it in 2013. But SHA-256 is a convention, not a requirement. Any collision-resistant hash function works.
Engram uses SHA3-256 (NIST FIPS 202) for three reasons:
-
Post-quantum collision resistance. SHA-256 has ~128-bit collision resistance classically, but quantum algorithms (Brassard-Hoyer-Tapp) reduce this to ~64-bit. SHA3-256 maintains 128-bit collision resistance even against quantum adversaries. For Merkle proofs used in selective disclosure (proving one artifact exists without revealing others), collision resistance is the property that matters.
-
Sponge construction. SHA-256 uses Merkle-Damgard construction, which is susceptible to length extension attacks. SHA3's sponge construction is immune by design. While length extension isn't a direct Merkle tree vulnerability, using a hash with fewer structural weaknesses reduces the attack surface.
-
Long-term audit trails. Engram's Merkle tree targets use cases with 10-30 year retention requirements — medical device compliance logs (FDA), regulatory audit records (NIST AI RMF), BCI session integrity (QIF). Choosing the stronger hash now costs nothing (SHA3-256 runs at comparable speed to SHA-256 in Rust with LTO). Upgrading later would require re-hashing every artifact in every tier.
The Merkle tree runs entirely in the Rust sidecar (engram-vault). SHA3-256 hashing, proof generation, proof verification, and HMAC root sealing all happen in compiled Rust with mlocked memory and zero core dumps. Python never handles hash computations for integrity-critical operations.
PQC Root Sealing
The Merkle root is sealed with HMAC-SHA3-256 using a key derived from the ML-KEM-768 (FIPS 203) shared secret via HKDF. This proves the root was computed by a process with access to the post-quantum key material — not just that the hashes are correct, but that the entity who computed them holds the PQC key.
ML-KEM-768 shared secret
→ HKDF-SHA3-256 key derivation
→ HMAC-SHA3-256(merkle_root)
→ root seal (32 bytes, constant-time verified)
Verification is constant-time in the Rust sidecar (prevents timing side-channel attacks on seal comparison).
The full PQC integrity stack:
| Layer | Algorithm | NIST Standard |
|---|---|---|
| Tree hashes | SHA3-256 | FIPS 202 |
| Root seal | HMAC-SHA3-256 | FIPS 198-1 + FIPS 202 |
| Seal key source | ML-KEM-768 + HKDF | FIPS 203 + SP 800-56C |
| Artifact encryption | AES-256-GCM | FIPS 197 + SP 800-38D |
| Key encapsulation | ML-KEM-768 + X25519 | FIPS 203 |
| All operations | Rust sidecar | mlocked, zero core dumps |
Least Privilege by Design
The tiered architecture isn't just compression. It's the principle of least privilege applied to AI context.
Traditional least privilege: a user only gets the permissions they need for the current task. Engram least privilege: the AI only gets the context fidelity it needs for the current task.
| Tier | Access Level | What the AI Sees | Analogy |
|---|---|---|---|
| Hot | Full access | Complete session content, every token | Root access — current working set |
| Warm | Summaries | Key decisions, outcomes, compressed context | Read access — recent history |
| Cold | Keywords only | Semantic index entries, topic references | Directory listing — you know it exists |
| Frozen | Explicit recall required | Nothing until engram recall is invoked |
Locked vault — must request access |
The constraint is real: the context window is finite. The enforcement is real: compression removes detail until you explicitly ask for it back. Old context doesn't linger at full fidelity. It compresses, indexes, and locks down — the same way a well-run system revokes elevated permissions after the task is done.
This means a compromised or misbehaving AI session can't silently access months of full-fidelity conversation history. It gets summaries and keywords. Full recall requires an explicit action (engram recall <path>) that can be logged, rate-limited, or gated behind encryption. The blast radius of any single session is bounded by its tier.
Add post-quantum encryption and the tiers become actual access barriers: even with disk access, frozen artifacts are ML-KEM-768 encrypted with per-artifact keys managed by a Rust sidecar that never lets private keys enter Python's address space.
FAQ — Merkle trees, Matryoshka embeddings, the Rust sidecar, PQC encryption, component architecture, performance benchmarks, and security review findings.
Table of Contents
- Quick Start
- How It Works
- Compression Pipeline
- Encryption
- Supported AI Assistants
- Architecture
- Security
- Use Cases
- Configuration
- CLI Reference
- Requirements
- Disclaimer
- AI Disclosure
- License
Quick Start
pip install bci-engram # install from PyPI
pip install bci-engram[embeddings] # with semantic search (recommended)
engram init # guided setup
engram run --dry-run # preview (safe, no changes)
engram run # compress + encrypt
engram search "authentication refactor" # search all tiers
engram context --query "auth patterns" # get context for your AI
engram recall <path> # decompress a cold artifact
engram status # tier distribution
Enable encryption
cd engram/sidecar && cargo build --release # build crypto sidecar
engram encrypt-setup # generate keys + store in Keychain
How It Works
Brain-inspired tiering. Artifacts move through tiers based on age and idle time:
| Transition | When | Pipeline | Ratio |
|---|---|---|---|
| Hot → Warm | 1 week old + 3 days idle | Minify JSON → zstd-3 | 4-5x |
| Warm → Cold | 1 month old + 2 weeks idle | Strip boilerplate → dict-trained zstd-9 | 8-12x |
| Cold → Frozen | 3 months old + 1 month idle | Columnar Parquet + dict + zstd-19 | 20-50x |
All thresholds configurable. Choose 2-tier (simple) or 4-tier (full) during setup.
Hybrid search + CoGraph. Every artifact is indexed with keywords, embeddings, LSH hashes, and HNSW graph entries before compression. Search uses a 6-layer retrieval stack: keyword lookup, LSH hash tables, HNSW nearest-neighbor graphs, Reciprocal Rank Fusion (BM25 + vector), optional reranking, and summary output. The index (~1 MB keywords + ~10 MB embeddings for 10K artifacts) is always loaded, never compressed. Search across all tiers without decompressing anything. Summaries load at 10-20% of token cost. Full recall on demand.
CoGraph — associative recall. When you recall one artifact, CoGraph automatically surfaces related ones. Layer 3 (co-occurrence) tracks which artifacts are recalled together using PPMI edge weighting. Layer 4 (spreading activation) propagates through the graph via weighted BFS — like the brain's spreading activation network (Collins & Loftus 1975). The graph lives entirely in the Rust sidecar's mlocked memory — never written to disk. Ephemeral by design: no persistence means no attack surface at rest. ~7 MB worst case at 10K artifacts.
Everything local. No data leaves your machine. No telemetry. No cloud dependency. Zero network calls. No third-party ML models. Keyword search + CoGraph handles retrieval with zero external dependencies. Optional embedding model available via pip install bci-engram[embeddings] for vector similarity — runs 100% local if installed.
What Engram encrypts today vs. what's in development
Currently encrypted: Old AI session logs (~/.claude/ subagent logs, compressed session data, debug traces). These are archival records — the LLM doesn't read raw session files for context. Instead, Engram's semantic index (keywords, summaries, tier metadata) stays loaded in memory during every session, so search works across all tiers without decrypting anything. Full content is available via engram recall when needed. Active session files (current history.jsonl, recent logs) stay in hot tier (plaintext) until they age past the configured threshold.
Working memory — in development. AI working memory files (persistent context the LLM reads across sessions) can't be encrypted yet. Hooks and context loaders read these at session start. Encrypting them breaks the context pipeline — the LLM can't load what it can't read, and decrypting dozens of files on every session start adds latency and auth friction.
This is the harder problem. The retrieval infrastructure is being built to solve it: CoGraph (co-occurrence tracking + spreading activation), the keyword inverted index, HNSW nearest-neighbor graphs, and Reciprocal Rank Fusion allow Engram to know what context is relevant without reading every file. Once the retrieval layer can reliably predict which memories the LLM needs for a given session, selective decryption becomes viable — decrypt only the few files that matter, not all of them.
This requires more testing to ensure the LLM doesn't lose critical context while memory files are encrypted at rest. The goal is encrypted-by-default working memory with retrieval-guided selective decryption. Not there yet.
Compression Pipeline
Not just "higher zstd levels" (that gets 3.2x to 3.8x). Each tier applies different data transformations:
- Warm: JSON minification strips 30-40% of whitespace before zstd-3
- Cold: Boilerplate stripping replaces repeated system prompts (2,000-5,000 tokens per session) with 64-byte hash refs. Dictionary trained on your session logs teaches zstd the shared schema. Only unique content gets compressed.
- Frozen: JSONL transposed to columnar Parquet.
rolecolumn (cardinality 2) → run-length encoded to nothing.timestampcolumn (monotonic) → delta encoded to nothing. Onlycontentcarries entropy. ClickHouse achieves 170x on logs with this approach.
Real results
4,564 artifacts | 2.6 GB raw | 11.62x ratio | 1.6 GB saved | 132K keywords indexed
Encryption
Not enabled by default. Engram works without it — compression, indexing, and context enhancement all function with zero encryption. To enable encryption, run engram encrypt-setup to generate ML-KEM-768 keypairs and store them in your OS credential vault (Keychain on macOS). Then set "encryption": {"enabled": true} in ~/.engram/config.json.
When enabled, uses ML-KEM-768 (NIST FIPS 203) in hybrid mode with X25519, using AES-256-GCM (NIST FIPS 197) for data encryption and HKDF-SHA256 (NIST SP 800-56C) for key derivation.
Crypto sidecar (engram-vault)
Private keys are handled by a compiled Rust binary — not Python. Keys never enter Python's address space, never touch disk, never appear in process args, never hit swap.
| Protection | How |
|---|---|
| Keys out of Python | Rust sidecar is the crypto boundary |
| Keys off disk | OS credential vault → mlock'd memory → in-process crypto → zeroed |
| Keys out of swap | mlockall() at startup |
| Keys out of core dumps | setrlimit(RLIMIT_CORE, 0) |
| Keys out of logs | Regex PII detector blocks secrets before writing |
| No env hijack | .env_clear() on all subprocess calls |
| Zeroize on all paths | Zeroizing<T> wrappers on all key material — errors, panics, normal exit |
Key material lifecycle
Private keys are never stored as environment variables and never passed over any wire or socket.
| Step | What happens |
|---|---|
| 1. Storage | Private keys live in your OS credential vault — macOS Keychain (encrypted by the OS, protected by login password/Touch ID), Windows Credential Manager (DPAPI-backed), or Linux libsecret/GNOME Keyring. The vault encrypts keys at rest using OS-level protections. |
| 2. Retrieval | The Rust sidecar calls the OS credential API (Security.framework on macOS) to read the key directly into its own mlock'd memory. No intermediate files, no pipes, no env vars. |
| 3. Usage | Key material stays in-process only. Used for ML-KEM-768 decapsulation + X25519 Diffie-Hellman + HKDF-SHA256 key derivation + AES-256-GCM encrypt/decrypt. All NIST-approved algorithms. |
| 4. Cleanup | Zeroizing<Vec<u8>> wrappers (RustCrypto zeroize crate) ensure deterministic zeroing on ALL exit paths — normal return, error return, and panic. Compiler barriers prevent the optimizer from eliding the zeroing. This satisfies NIST SP 800-57 Part 1 Section 8.3 key destruction requirements. |
| 5. Never | Touches disk. Enters Python. Appears in process args. Passes through env vars. Goes over any wire or socket. Gets logged. Hits swap (mlockall). Appears in core dumps (RLIMIT_CORE=0). |
The Python side only ever sees the public key (for encryption). The sidecar is the crypto boundary.
Two modes
| Simple | Envelope | |
|---|---|---|
| Keys | One for all tiers | Per-tier keypairs |
| Per-artifact isolation | No | Yes — unique 256-bit DEK each |
| Compromise radius | All data | One tier or one artifact |
| Key rotation | Re-encrypt all (O(data)) | Re-wrap headers (O(metadata)) |
Why post-quantum now
NIST IR 8547 (draft, Nov 2024): RSA-2048/P-256 deprecated by 2030, disallowed by 2035. "Harvest now, decrypt later" is a real threat. Starting with classical crypto means migrating before 2030. Start with PQ and you're done.
Supported AI Assistants
Tested and verified with Claude Code. Other assistants are auto-detected and scanned but have not been validated end-to-end. Use at your own discretion — Engram is being actively improved as testing expands and efficiency is refined over time. Community testing and feedback welcome.
| Assistant | Auto-detected | Tested |
|---|---|---|
| Claude Code | 18 locations: sessions, subagents, memory, debug, plans, tasks | Yes |
| OpenClaw | MEMORY.md, daily notes, sessions, context | Scan only |
| ChatGPT | Desktop app cache | Scan only |
| Cursor | Conversation logs | Scan only |
| GitHub Copilot | Configuration cache | Scan only |
| Any AI tool | Add any directory to config.json |
— |
Architecture
For the full detailed walkthrough of how every component works — registration, tier transitions, compression pipeline, hybrid search retrieval stack, HNSW vector index, embedding architecture, recall flow, encryption lifecycle, lookup tables, and index encryption — see docs/ARCHITECTURE.md.
KV cache reduction through lookup tables. The same insight that makes DeepSeek-V2 fast applies here: don't recompute what you can store compressed and look up. Every retrieval path in Engram uses precomputed structures — inverted keyword indexes (O(1) per keyword), LSH hash tables (O(1) approximate nearest neighbor), HNSW graphs (O(log n) exact nearest neighbor), Product Quantization codebooks (192x embedding compression), and trained compression dictionaries. The net effect: 10,000-artifact search completes in under 50ms without decompressing a single file.
File tree
engram/
├── sidecar/ # Rust crypto + Merkle-Index + CoGraph sidecar (670 KB binary)
│ ├── Cargo.toml
│ └── src/
│ ├── main.rs # Protocol handler, security hardening (mlockall, zero core dumps)
│ ├── merkle.rs # SHA3-256 Merkle tree + inverted keyword index + HMAC seal
│ ├── cograph.rs # CoGraph: co-occurrence tracking + spreading activation (Layer 3+4)
│ ├── simhash.rs # SimHash fingerprinting for semantic deduplication
│ ├── crypto.rs # ML-KEM-768 + X25519 + AES-256-GCM (NIST-only)
│ ├── keystore.rs # Cross-platform key management
│ └── keychain.rs # macOS Security.framework integration
├── src/ # Python: orchestration, CLI, search
│ ├── engine.py # Tiering engine (scan → index → tier → recall)
│ ├── pipeline.py # Compression pipeline (minify → strip → dict → Parquet)
│ ├── context.py # Semantic index + context builder
│ ├── session_parser.py # Conversation content extraction from AI session files
│ ├── predictor.py # Matryoshka cascading predictive context loader
│ ├── spatial.py # Spatial memory extension (Spot integration)
│ ├── embeddings.py # Matryoshka tiered embeddings (384/256/128/64-dim)
│ ├── vector_index.py # HNSW nearest-neighbor per tier
│ ├── hybrid_search.py # Reciprocal rank fusion (keyword + vector)
│ ├── lookup_tables.py # LSH hash tables + Product Quantization codebook
│ ├── envelope.py # Per-artifact envelope encryption (DEK per file)
│ ├── cograph.py # CoGraph client (co-occurrence + spreading activation)
│ ├── vault.py # Python client for Rust sidecar
│ ├── index_crypto.py # Index bundle encryption (lock/unlock)
│ ├── encryption.py # Encryption orchestration
│ ├── compressor.py # zstd streaming compress/decompress
│ ├── config.py # Config schema + sensitive dir blocklist
│ ├── metadata.py # Artifact registry + SHA-256 integrity
│ ├── scanner.py # AI assistant artifact auto-detection (Claude, OpenClaw, ChatGPT, Cursor, Copilot)
│ ├── setup.py # Guided/interactive/auto setup wizard
│ ├── audit.py # Audit logger with PII/secret detection
│ ├── fileutil.py # Atomic writes, path containment, hashing
│ └── cli.py # CLI entry point (14 commands)
├── docs/
│ ├── FAQ.md # 30+ questions across 8 topics
│ ├── ARCHITECTURE.md # Full technical deep dive
│ ├── ARCHITECTURE-PROPOSAL.md # v2 architecture with measured benchmarks
│ ├── ENGRAM-V2-ARCHITECTURE.md # Brain-informed design (quorum-reviewed)
│ ├── MERKLE-INDEX-SPEC.md # Merkle-as-Index technical spec
│ ├── SPATIAL-EXTENSION.md # Spot spatial memory integration
│ ├── KEY-STORAGE-GUIDE.md # Key management (Keychain, Vault, YubiKey)
│ └── blog-post.md # Launch blog post
├── tests/ # 189 tests (171 Python + 18 Rust)
│ ├── test_activation.py # CoGraph: unit, mock, sidecar integration (30 tests)
│ ├── test_merkle.py # Merkle tree via Rust sidecar
│ ├── test_spatial.py # Spatial memory extension
│ ├── benchmark_retrieval.py # Retrieval quality evaluation (Recall@k, MRR)
│ ├── test_engine.py
│ ├── test_compressor.py
│ ├── test_envelope.py
│ ├── test_metadata.py
│ ├── test_pipeline.py
│ ├── test_embeddings.py
│ └── test_lookup_tables.py
├── skills/engram/
│ └── SKILL.md # Claude Code skill (13 use cases)
├── hooks/
│ └── hooks.json # SessionStart + PreCompact hooks
├── archived/
│ └── merkle_python_reference.py # Original Python Merkle (replaced by Rust)
├── marketplace.json # Distribution metadata
├── pyproject.toml # Package config (pip installable)
├── LICENSE # MIT
└── README.md
Documentation
| Document | Contents |
|---|---|
| FAQ | 30+ questions: Merkle trees, Matryoshka embeddings, sidecar, PQC, performance, security |
| Architecture | Full technical deep dive: registration, tiers, compression, search, encryption, audit |
| v2 Architecture | Brain-informed design, quorum-reviewed: verified summaries, section recall, CoGraph |
| Merkle-Index Spec | Merkle tree as search accelerator: measured 400x lookup improvement, sidecar protocol |
| Architecture Proposal | Full v2 proposal: system diagrams, data flows, performance model, implementation strategy |
| Spatial Extension | Spot integration: tiered spatial memory, NSP peer sharing, QIF AI0 stack |
| Key Storage Guide | Key management: Keychain, Vault, YubiKey, environment variables |
Security
- 11 rounds of security review (red team, crypto, GRC, supply chain, OWASP, embedding injection, index encryption, lookup table integrity, hybrid search fuzzing, CoGraph audit)
- Rust crypto sidecar — keys never in Python
- PQ encryption (ML-KEM-768 + X25519 hybrid)
- Per-artifact DEKs in envelope mode
- Audit logging with regex PII/secret detection
- SHA-256 integrity verification (quantum-safe: Grover reduces to 128-bit equivalent, well above NIST minimum per SP 800-57)
- Symlink protection, path containment, sensitive directory blocklist
- Input validation (newlines/nulls rejected in protocol)
- Core dumps disabled, memory locked, environment cleared
- No
shell=Trueanywhere - Embedding model supply chain verification (SHA-256 checksum on load)
Third-party model verification
Engram uses one third-party model for semantic search (optional):
| Property | Value |
|---|---|
| Model | sentence-transformers/all-MiniLM-L6-v2 |
| License | Apache 2.0 |
| Size | 86.7 MB (model.safetensors) |
| Parameters | 22.7M |
| SHA-256 | 53aa51172d142c89d9012cce15ae4d6cc0ca6895895114379cacb4fab128d9db |
| Downloads | 207M+ (widely vetted) |
| Runs | 100% local, in-process, CPU. No API calls. |
The checksum is verified on every model load. If the cached weights don't match the expected hash, Engram refuses to load the model and logs a security error. This detects supply chain tampering via compromised HuggingFace cache, typosquatted packages, or modified downloads.
Without the embedding model, keyword search still works. The model is an optional dependency (pip install engram[embeddings]).
Threat model
Protects against: data at rest on stolen devices, harvest-now-decrypt-later, unauthorized disk reads, forensics on decommissioned hardware.
Does NOT protect against: compromised process with same-UID ptrace, attacker with your Keychain password, trojanized engram-vault binary, root compromise on non-Secure-Enclave hardware.
Recommendations
See the full security recommendations section for the 10-point hardening guide (shared server isolation, cron lockdown, key compromise response, secure uninstall, etc.).
Use Cases
Solo dev: 6 months of sessions → compressed from 200 MB to 15 MB. Search past decisions without re-explaining.
Security researcher: Sessions contain exploits and disclosure timelines. PQ envelope encryption. Touch ID on recall. Private keys in Keychain-backed Keychain.
Team server: Shared ~/.claude/. Per-user encryption keys. Semantic index lets everyone search. Content isolation via per-tier keypairs.
Multi-project: Point Engram at 5 project memory dirs. Cross-project search finds the answer regardless of which project it was in.
Spatial memory (Spot): Engram extends beyond AI context to spatial data — LiDAR geometry, walking routes, indoor maps, bench/bathroom locations. The same tiered compression that handles conversation history handles spatial memory for Spot, a LiDAR navigation app. Routes compress from hot (live LiDAR) to frozen (Parquet-archived geometry). Peer-to-peer spatial sharing between devices uses NSP-wrapped payloads with Merkle proofs for selective verification. See docs/SPATIAL-EXTENSION.md.
Configuration
All thresholds configurable in ~/.engram/config.json:
{
"tier_policy": {
"hot_to_warm_age_hours": 48,
"warm_to_cold_age_hours": 336,
"cold_to_frozen_age_hours": 2160
},
"encryption": {
"enabled": true,
"envelope_mode": true,
"warm_pubkey": "<hex from KEYGEN>",
"warm_private_source": "keychain:engram:warm-key"
},
"audit_log": false
}
Setup modes: engram init (guided), --mode interactive (pick locations), --mode auto (no prompts).
CLI Reference
| Command | What |
|---|---|
engram init |
Guided setup with tier choice and encryption |
engram scan |
Discover artifacts |
engram run |
Execute tier transitions |
engram run --dry-run |
Preview without changes |
engram status |
Tier distribution and stats |
engram search <query> |
Search across all tiers |
engram context --query <q> |
Budget-optimized context block |
engram recall <path> |
Decompress artifact to hot |
engram reindex |
Rebuild semantic index |
engram verify |
SHA-256 integrity check |
engram encrypt-setup |
Configure encryption |
Requirements
pip install bci-engram # core (compression + indexing)
pip install bci-engram[embeddings] # + semantic search (recommended)
- Python 3.10+
zstandard>= 0.19.0 (auto-installed)pyarrow>= 14.0.0 (auto-installed, for frozen Parquet tier)sentence-transformers>= 2.2.0 (optional, for semantic search) —pip install bci-engram[embeddings]- Rust toolchain (optional, to build the crypto sidecar) —
brew install rust(macOS) or see rustup.rs
Disclaimer
USE AT YOUR OWN RISK. This software is provided as-is under the MIT license with no warranty of any kind, express or implied, including but not limited to warranties of merchantability, fitness for a particular purpose, or non-infringement.
Data handling. Engram compresses and optionally encrypts files on your local filesystem. Compression transformations (JSON minification, boilerplate stripping) are lossy for whitespace. Full semantic content is preserved and recoverable via engram recall, but byte-for-byte identity with the original file is not guaranteed after tiering.
Encryption limitations. If you lose your private key, encrypted data is unrecoverable. There is no backdoor, no recovery mechanism, no reset. The encryption is only as strong as your key management. The Rust crypto sidecar uses the ml-kem (v0.2, RustCrypto), aes-gcm (v0.10, RustCrypto), and x25519-dalek (v2) crates. These crates have not received a combined independent audit specific to this integration.
Not a certified security product. Engram has been reviewed through multiple rounds of automated security analysis but has not been formally audited by an independent security firm, is not FIPS-validated, and does not hold any compliance certifications. It is not a substitute for an HSM, a certified KMS, or a CMVP-validated crypto module.
No liability for data loss. The author is not responsible for data loss, corruption, unauthorized access, or any damages resulting from use of this software. Back up your data before enabling compression or encryption. Test engram recall before relying on it for critical data.
Regulatory compliance is your responsibility. Engram provides technical controls (encryption, audit logging, access separation) but does not constitute compliance with GDPR, HIPAA, SOC 2, PCI-DSS, or any other framework. Consult your compliance team.
AI Disclosure
Built with AI assistance (Claude Opus 4.6, Anthropic). AI was used for code generation, security review, compression research, documentation, and test generation. All code reviewed by the author. All factual claims fact-checked against primary sources. The author takes full responsibility for all published content. Every git commit includes Co-Authored-By for transparency.
License
MIT
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file bci_engram-1.2.0.tar.gz.
File metadata
- Download URL: bci_engram-1.2.0.tar.gz
- Upload date:
- Size: 154.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
a4b2bedf1df53603491aeab3c609a5288eb7382d736fe6049d3bddfde025ee07
|
|
| MD5 |
2a731ca16df316377021b915f1ef4b4b
|
|
| BLAKE2b-256 |
4d09198e7fb85e4b9b956ea237ba2bfadbd2b2e97a622b0a30a558167826bf33
|
Provenance
The following attestation bundles were made for bci_engram-1.2.0.tar.gz:
Publisher:
publish.yml on qinnovates/engram
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
bci_engram-1.2.0.tar.gz -
Subject digest:
a4b2bedf1df53603491aeab3c609a5288eb7382d736fe6049d3bddfde025ee07 - Sigstore transparency entry: 1155044136
- Sigstore integration time:
-
Permalink:
qinnovates/engram@1b2ebb28d88e8ae468b0fab5cc161164de807670 -
Branch / Tag:
refs/tags/v1.2.0 - Owner: https://github.com/qinnovates
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@1b2ebb28d88e8ae468b0fab5cc161164de807670 -
Trigger Event:
push
-
Statement type:
File details
Details for the file bci_engram-1.2.0-py3-none-any.whl.
File metadata
- Download URL: bci_engram-1.2.0-py3-none-any.whl
- Upload date:
- Size: 143.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
ae4e3178e51d9300a1d0a4f3bf538c3f7ae6cc67e8d0cba67e398f8470e94340
|
|
| MD5 |
5c3aba3aa9b04950a085c49e78213932
|
|
| BLAKE2b-256 |
39b7fe0dbb5bee8f451feedcd2f5c6c7887e3bb1c0ab3f956b057c18231c9779
|
Provenance
The following attestation bundles were made for bci_engram-1.2.0-py3-none-any.whl:
Publisher:
publish.yml on qinnovates/engram
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
bci_engram-1.2.0-py3-none-any.whl -
Subject digest:
ae4e3178e51d9300a1d0a4f3bf538c3f7ae6cc67e8d0cba67e398f8470e94340 - Sigstore transparency entry: 1155044137
- Sigstore integration time:
-
Permalink:
qinnovates/engram@1b2ebb28d88e8ae468b0fab5cc161164de807670 -
Branch / Tag:
refs/tags/v1.2.0 - Owner: https://github.com/qinnovates
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@1b2ebb28d88e8ae468b0fab5cc161164de807670 -
Trigger Event:
push
-
Statement type: