Emotional memory library for LLMs based on Affective Field Theory
Project description
emotional_memory
Emotional memory for LLMs based on Affective Field Theory (AFT) — a 5-layer model that encodes not just what happened, but how it felt, how that feeling was moving, and what mood colored the moment.
Installation
uv pip install emotional-memory
uv pip install "emotional-memory[sentence-transformers]" # real semantic embeddings (recommended)
uv pip install "emotional-memory[sqlite]" # SQLite persistence via sqlite-vec
uv pip install "emotional-memory[viz]" # matplotlib visualization
uv pip install "emotional-memory[dotenv]" # .env file loading via python-dotenv
For development:
git clone https://github.com/gianlucamazza/emotional-memory
cd emotional-memory
make install
# optional local demo stack:
make install-demo
Quickstart
uv pip install "emotional-memory[sentence-transformers]"
from emotional_memory import EmotionalMemory, InMemoryStore, CoreAffect
from emotional_memory.embedders import SentenceTransformerEmbedder
em = EmotionalMemory(
store=InMemoryStore(),
embedder=SentenceTransformerEmbedder(), # all-MiniLM-L6-v2 by default
)
# Set current emotional state
em.set_affect(CoreAffect(valence=0.8, arousal=0.6))
# Encode memories — each one captures the full affective context
em.encode("Just shipped the feature after three hard weeks.")
em.encode("Team celebration in the office.", metadata={"source": "slack"})
# Retrieve — ranked by semantic relevance AND emotional congruence
results = em.retrieve("difficult project success", top_k=3)
for mem in results:
print(mem.content, mem.tag.core_affect)
# Or inspect why a memory ranked where it did
explained = em.retrieve_with_explanations("difficult project success", top_k=1)
top = explained[0]
print(top.score)
print(top.breakdown.raw_signals)
Bring your own embedder — any object with .embed(text) -> list[float] works:
class MyEmbedder:
def embed(self, text: str) -> list[float]: ...
def embed_batch(self, texts: list[str]) -> list[list[float]]: ...
Or subclass SequentialEmbedder and implement only embed() — embed_batch() is provided.
Async
import asyncio
from emotional_memory import EmotionalMemory, InMemoryStore, as_async
from emotional_memory.embedders import SentenceTransformerEmbedder
sync_em = EmotionalMemory(store=InMemoryStore(), embedder=SentenceTransformerEmbedder())
em = as_async(sync_em) # wraps sync components with asyncio.to_thread bridges
async def main():
await em.encode("Meeting went surprisingly well today.")
results = await em.retrieve("work meeting", top_k=3)
asyncio.run(main())
For native async embedders or stores, construct AsyncEmotionalMemory directly with
SyncToAsyncEmbedder, SyncToAsyncStore, or your own AsyncEmbedder/AsyncMemoryStore.
Affective Field Theory
AFT models emotion as a field — distributed, dynamic, multi-layer — rather than a discrete label or a single coordinate. Five layers are captured at encoding time:
| Layer | Model | What it captures |
|---|---|---|
| CoreAffect | Barrett/Russell circumplex | Continuous (valence, arousal) — the emotional substrate |
| AffectiveMomentum | Spinoza — affect as transition | Velocity and acceleration of affect change |
| MoodField | Heidegger — Stimmung as attunement | Slow-moving global mood with inertia (EMA) |
| AppraisalVector | Scherer/Lazarus/Stoics | Emotion derived from evaluation: novelty, goal-relevance, coping, norm-congruence, self-relevance |
| ResonanceLinks | Aristotle/Hume/Bower/Collins & Loftus/Hebb | Associative bidirectional graph: semantic, emotional, temporal, causal, contrastive links; multi-hop spreading activation + Hebbian co-retrieval strengthening |
Full theoretical foundations: docs/research/
API Overview
EmotionalMemory
em = EmotionalMemory(
store: MemoryStore,
embedder: Embedder,
appraisal_engine: AppraisalEngine | None = None, # optional: auto-appraise via LLM
config: EmotionalMemoryConfig | None = None,
)
| Method | Description |
|---|---|
encode(content, appraisal=None, metadata=None) -> Memory |
Encode content with full AFT pipeline |
observe(content, appraisal=None, metadata=None) -> EmotionalTag |
Update affective state without storing a retrievable memory |
encode_batch(contents, metadata=None) -> list[Memory] |
Batch encode with embed_batch(), per-item appraisal |
retrieve(query, top_k=5) -> list[Memory] |
Emotionally-weighted retrieval + reconsolidation |
retrieve_with_explanations(query, top_k=5) -> list[RetrievalExplanation] |
Same retrieval pipeline plus a structured score breakdown |
elaborate(memory_id) -> Memory | None |
Run full appraisal on a fast-path (pending_appraisal=True) memory and blend core_affect |
elaborate_pending() -> list[Memory] |
Elaborate all pending fast-path memories in one call |
delete(memory_id) |
Remove a memory from the store |
get(memory_id) -> Memory | None |
Look up a single memory by ID |
list_all() -> list[Memory] |
Return all stored memories |
len(engine) -> int |
Number of memories in the store |
prune(threshold=0.05) -> int |
Delete memories below effective strength threshold; returns count removed |
export_memories() -> list[dict] |
Serialise all memories to JSON-safe dicts (backup / migration) |
import_memories(data, overwrite=False) -> int |
Restore from export_memories() output; returns count written |
get_state() -> AffectiveState |
Current affective state (read-only copy) |
set_affect(core_affect) |
Manually inject a CoreAffect |
reset_state() |
Reset runtime affective state to the initial baseline |
save_state() -> dict |
Serialise affective state for persistence |
load_state(data) |
Restore previously saved affective state |
get_current_mood(now=None) -> MoodField |
Read-only mood with time regression |
close() |
Release store resources (e.g. SQLite connection); also via with |
Both engines support context managers for automatic resource cleanup:
with EmotionalMemory(store=SQLiteStore("mem.db"), embedder=MyEmbedder()) as em:
em.encode("Session start")
results = em.retrieve("relevant context")
# SQLiteStore.close() called automatically
Use retrieve() for normal recall paths. Use retrieve_with_explanations() when
you need the ranking-time decomposition (semantic, mood, affect, momentum,
recency, resonance) for debugging, evaluation, or UI inspection.
AsyncEmotionalMemory
Same method signatures as EmotionalMemory. Coroutines: encode, observe, encode_batch, retrieve,
retrieve_with_explanations, elaborate, elaborate_pending, delete, get, list_all,
count, prune, export_memories, import_memories, close.
State accessors (get_state, set_affect, reset_state, save_state, load_state, get_current_mood)
remain synchronous.
Supports async with for automatic resource cleanup:
async with AsyncEmotionalMemory(store=..., embedder=...) as em:
await em.encode("Session start")
results = await em.retrieve("relevant context")
from emotional_memory import AsyncEmotionalMemory, SyncToAsyncEmbedder, SyncToAsyncStore
Bridge adapters: SyncToAsyncEmbedder, SyncToAsyncStore, SyncToAsyncAppraisalEngine wrap
any sync implementation. SyncToAsyncStore also proxies close() to the underlying store.
as_async(engine) wraps a complete EmotionalMemory in one call.
Key config classes
EmotionalMemoryConfig— top-level config (decay, retrieval, resonance, mood_alpha, mood_decay)RetrievalConfig— weights, APE threshold, reconsolidation learning rateResonanceConfig— similarity threshold, max links, semantic/emotional/temporal weights, candidate multiplier,propagation_hops,hebbian_increment, configurable link-classification thresholdsDecayConfig— power-law decay parameters, arousal modulation, floor valuesMoodDecayConfig— time-based mood regression (half-life, inertia scale, baselines)AdaptiveWeightsConfig— smooth mood-adaptive retrieval weight tuning (sigmoid/Gaussian gates)LLMAppraisalConfig— LLM appraisal engine settings (system prompt, cache size, fallback behaviour)
Interfaces (bring your own)
If your embedder has no native batching, subclass SequentialEmbedder — embed_batch() is
provided automatically:
from emotional_memory import SequentialEmbedder
class MyEmbedder(SequentialEmbedder):
def embed(self, text: str) -> list[float]:
return my_model.encode(text).tolist()
Otherwise implement the full Embedder protocol:
class Embedder(Protocol):
def embed(self, text: str) -> list[float]: ...
def embed_batch(self, texts: list[str]) -> list[list[float]]: ...
class MemoryStore(Protocol):
def save(self, memory: Memory) -> None: ...
def get(self, memory_id: str) -> Memory | None: ...
def update(self, memory: Memory) -> None: ...
def delete(self, memory_id: str) -> None: ...
def list_all(self) -> list[Memory]: ...
def search_by_embedding(self, embedding: list[float], top_k: int) -> list[Memory]: ...
def __len__(self) -> int: ...
Async variants (AsyncEmbedder, AsyncMemoryStore, AsyncAppraisalEngine) are defined in
interfaces_async.py. AsyncMemoryStore uses count() -> int instead of __len__ since
dunder methods cannot be coroutines.
Stores included:
InMemoryStore— dict-backed, brute-force cosine search (no extra deps)SQLiteStore— persistent SQLite + sqlite-vec ANN search (uv pip install "emotional-memory[sqlite]")
Appraisal Engines
class AppraisalEngine(Protocol):
def appraise(self, event_text: str, context: dict | None = None) -> AppraisalVector: ...
Pass an appraisal_engine to EmotionalMemory to auto-generate AppraisalVector during encode.
LLMAppraisalEngine — wrap any LLM SDK in a single callable:
from emotional_memory import LLMAppraisalEngine
def my_llm(prompt: str, json_schema: dict) -> str:
# call openai / anthropic / local model here
return response_text
engine = LLMAppraisalEngine(llm=my_llm)
em = EmotionalMemory(store=..., embedder=..., appraisal_engine=engine)
KeywordAppraisalEngine — regex-based fallback, zero external dependencies, ships with
default rules covering success, failure, novelty, danger, and social norms:
from emotional_memory import KeywordAppraisalEngine
engine = KeywordAppraisalEngine() # or pass custom KeywordRule list
Visualization
The optional viz extra provides 8 plotting functions for inspecting and presenting the model's internals. Each function accepts an optional ax parameter for subplot composition and returns a matplotlib.Figure.
from emotional_memory.visualization import plot_circumplex, plot_decay_curves
Valence-Arousal Circumplex
Memories plotted on Russell's (1980) 2D circumplex, colored by consolidation strength.
Decay Curves (ACT-R Power Law)
Family of curves showing how arousal (McGaugh 2004) and retrieval count (spacing effect) modulate memory decay.
Yerkes-Dodson Inverted-U
Consolidation strength peaks near effective arousal 0.7, then drops — the classic Yerkes-Dodson curve.
6-Signal Retrieval Breakdown
Radar chart of the six retrieval signals: semantic similarity, mood congruence, affect proximity, momentum alignment, recency, and resonance boost.
Mood Field Evolution
Time series of valence, arousal, and dominance with dashed baselines showing the regression attractors.
Adaptive Retrieval Weights
Heatmap showing how retrieval weights shift across different mood states (valence x arousal grid).
Resonance Network
Directed graph with memories as nodes and edges colored by link type (semantic, emotional, temporal, causal, contrastive).
Appraisal Radar (Scherer CPM)
Spider chart of the 5 Stimulus Evaluation Check dimensions.
Generating images
make docs-images # regenerate all PNGs in docs/images/
Comparison with Existing Systems
| emotional-memory (AFT) | Mem0 | Letta | Zep | LangChain Memory | |
|---|---|---|---|---|---|
| License | MIT | Apache 2.0 | Apache 2.0 | Apache 2.0 | MIT |
| Persistence | InMemory / SQLite | Qdrant, Chroma, Pinecone, PG, MongoDB | PostgreSQL / SQLite | Neo4j (self-hosted) / Cloud | In-memory / custom |
| BYO embedder | ✅ any Embedder protocol |
✅ (OpenAI default) | ⚠️ partial | ⚠️ partial | ✅ |
| Emotion model | ✅ 5-layer AFT (valence, arousal, mood, appraisal, resonance) | ❌ | ❌ | ❌ | ❌ |
| Reconsolidation | ✅ APE-gated lability window | ✅ auto update/remove | ✅ tool-call edit | ✅ edge invalidation | ❌ |
| Persistent mood state | ✅ MoodField (Heidegger EMA) | ❌ | ❌ | ❌ | ❌ |
| LLM-agnostic | ✅ | ✅ | ✅ | ✅ | ✅ |
| LangChain integration | ✅ EmotionalMemoryChatHistory |
✅ official | ✅ tools interop | ✅ ZepVectorStore | ✅ native |
| Public benchmark | ✅ 126 fidelity test cases (intra-theory) | ✅ LoCoMo, LongMemEval, BEAM | ✅ LoCoMo, DMR | ✅ DMR, LongMemEval | ❌ |
| Codebase size | ~4.8k LOC (src/) | >50k LOC | >50k LOC | >50k LOC | ~5k LOC |
Key differentiator: emotional-memory is the only system with an explicit multi-layer emotional model. Every other system treats memory as semantically-indexed text with no affective state — no mood congruence in retrieval, no reconsolidation triggered by affective prediction error, no momentum or appraisal.
Benchmark caveat: AFT fidelity tests validate psychological invariants (intra-theory); Mem0/Letta/Zep benchmarks measure QA recall on dialog datasets (inter-system). A cross-system comparison on affect-labeled retrieval tasks is in progress (v0.6).
Benchmarks
Psychological fidelity (126 parametrized test cases, 20 phenomena)
The library validates 20 phenomena from the affective science literature via 126 parametrized test cases (run pytest --collect-only benchmarks/fidelity/ to enumerate them):
| Phenomenon | Reference | Cases | Test file |
|---|---|---|---|
| Mood-congruent recall | Bower 1981 | 3 | test_mood_congruent.py |
| Emotional enhancement | Cahill & McGaugh 1995 | 3 | test_emotional_enhancement.py |
| Yerkes-Dodson inverted-U | Yerkes & Dodson 1908 | 12 | test_yerkes_dodson.py |
| Spacing effect | Ebbinghaus 1885 | 7 | test_spacing_effect.py |
| Arousal floor | McGaugh 2004 | 7 | test_arousal_floor.py |
| Reconsolidation (APE) | Nader & Schiller 2000 | 5 | test_reconsolidation.py |
| State-dependent retrieval | Godden & Baddeley 1975 | 3 | test_state_dependent.py |
| Affective momentum | Spinoza, Ethics III | 9 | test_momentum.py |
| Mood-adaptive weights | Heidegger, Being & Time §29 | 14 | test_mood_adaptive.py |
| Appraisal-to-affect mapping | Scherer CPM 2009 | 11 | test_appraisal_affect.py |
| Spreading activation | Collins & Loftus 1975 | 5 | test_spreading_activation.py |
| Hebbian co-retrieval strengthening | Hebb 1949 | 4 | test_hebbian_strengthening.py |
| ACT-R power-law decay | Anderson 1983 / McGaugh 2004 | 5 | test_decay_power_law.py |
| PAD dominance | Mehrabian & Russell 1974 | 8 | test_pad_dominance.py |
| Emotional retrieval vs. cosine | Bower 1981 / Russell 1980 / Nader 2000 | 3 | test_emotional_vs_cosine.py |
| Design gap regression | (various) | 3 | test_design_gaps.py |
| Dual-path encoding | LeDoux 1996 | 6 | test_dual_path_encoding.py |
| Emotion categorization | Plutchik 1980 | 10 | test_emotion_categorization.py |
| Affective prediction error | Schultz 1997 / Pearce-Hall 1980 | 5 | test_prediction_error.py |
| APE-gated reconsolidation window | Nader & Schiller 2000 | 3 | test_reconsolidation_window.py |
Run with: make bench-fidelity
Performance (hash-based embedder, InMemoryStore)
| Operation | N | Mean | OPS |
|---|---|---|---|
| Encode (single) | 1 | 1.7 ms | 590/s |
| Encode (batch of 100) | 100 | 9.9 ms/op | 101/s |
| Encode w/ resonance graph | 500 | 4.0 ms | 250/s |
| Retrieve top-5 | 100 | ~2 ms | ~500/s |
| Retrieve top-5 | 1 000 | ~12 ms | ~85/s |
| Retrieve top-5 | 10 000 | ~120 ms | ~8/s |
| Retrieve (top-k 1–25) | 1 000 | 10–18 ms | 55–100/s |
| Retrieve + reconsolidation | 200 | 2.6 ms | 385/s |
InMemoryStore.search_by_embedding uses vectorized matrix multiplication (numpy),
making retrieval O(n · d) in a single batch rather than n individual cosine calls.
Retrieval uses two-pass scoring (spreading activation); when no resonance links are
active the second pass is skipped. For stores > 10 000 memories, use SQLiteStore
(sqlite-vec ANN) or a vector database implementing the MemoryStore protocol.
Run with: make bench-perf
Appraisal quality (LLM prompt validation)
15 natural-language phrases with expected directional outcomes against Scherer's 5 dimensions:
| Phrase category | Key assertions |
|---|---|
| Personal loss ("I got fired") | goal_relevance < -0.2, coping_potential < 0.6 |
| Achievement ("Got promoted") | goal_relevance > 0.2, norm_congruence > 0.0 |
| Moral violation ("Coworker stole credit") | norm_congruence < -0.2, goal_relevance < 0.0 |
| Grief, danger, betrayal, relief, … | dimension-specific directional bounds |
Assertions use wide bands (e.g. > 0.3, < -0.2) and evaluate the median over 3 LLM calls to tolerate non-determinism. Designed to catch systematic prompt regressions, not exact calibration.
Run with: EMOTIONAL_MEMORY_LLM_API_KEY=... make bench-appraisal
Works with any OpenAI-compatible endpoint (Ollama, vLLM, LiteLLM, …) via EMOTIONAL_MEMORY_LLM_BASE_URL.
LangChain integration
EmotionalMemoryChatHistory is a drop-in replacement for any LangChain chat history object.
It backs the transcript with an EmotionalMemory instance so the affective state evolves
naturally as the conversation unfolds, while letting you control which messages become
retrievable memories.
uv pip install "emotional-memory[langchain,sentence-transformers]"
from emotional_memory import EmotionalMemory, InMemoryStore
from emotional_memory.embedders import SentenceTransformerEmbedder
from emotional_memory.integrations import (
EmotionalMemoryChatHistory,
recommended_conversation_policy,
)
em = EmotionalMemory(
store=InMemoryStore(),
embedder=SentenceTransformerEmbedder(),
)
history = EmotionalMemoryChatHistory(em, message_policy=recommended_conversation_policy)
# Works anywhere BaseChatMessageHistory is accepted:
history.add_user_message("I'm anxious about the deadline.")
history.add_ai_message("Let's break the work into smaller steps.")
print(history.messages) # [HumanMessage(...), AIMessage(...)]
# The underlying engine has tracked the affective state:
state = em.get_state()
print(f"valence={state.core_affect.valence:.2f} arousal={state.core_affect.arousal:.2f}")
With recommended_conversation_policy, user messages become retrievable memories, assistant
messages update affective state without being stored, and control commands such as
recall ... are ignored by retrieval. The adapter uses dependency injection — pass a
fully-configured EmotionalMemory so you control the store backend and embedder.
clear() removes stored memories, clears the transcript, and resets affective state.
Logging
The library uses the standard logging module. Enable debug output to trace the full pipeline:
import logging
logging.basicConfig(level=logging.DEBUG)
# or just for emotional_memory:
logging.getLogger("emotional_memory").setLevel(logging.DEBUG)
Debug events include: encode start/stored/resonance, retrieve start/done, reconsolidation triggers, LLM appraisal cache hits, and fallback activations.
Examples
The examples/ directory contains runnable scripts covering the full API.
All scripts are self-contained and use a deterministic HashEmbedder so they run without
any ML dependencies.
| Script | Description | Extra |
|---|---|---|
basic_usage.py |
Encode/retrieve, reconsolidation, resonance links | — |
advanced_config.py |
ACT-R decay, mood regression, adaptive weights | — |
appraisal_engines.py |
Keyword, static, and custom appraisal rules | — |
reconsolidation.py |
Two-retrieval lability window (Nader & Schiller 2000) | — |
async_usage.py |
as_async(), SyncToAsync* adapters, encode_batch |
— |
persistence.py |
SQLiteStore, save/load state, export/import, prune | [sqlite] |
emotional_journal.py |
Multi-session journaling with full lifecycle | [sqlite] |
llm_appraisal.py |
LLM-backed appraisal via OpenAI-compatible API | openai |
httpx_llm_integration.py |
httpx LLMCallable, .env config, 7 API deep-dives |
httpx |
sentence_transformers_embedder.py |
SequentialEmbedder with real embeddings |
sentence-transformers |
visualization.py |
All 8 matplotlib plot types | [viz] |
resonance_network.py |
Resonance graph and link-type distribution | [viz] |
retrieval_signals.py |
6-signal decomposition, radar chart, weight heatmap | [viz] |
Run any script: uv run python examples/<script>.py
Development
make check # lint + typecheck + test
make cov # tests with branch coverage report
make bench # fidelity + performance benchmarks
# Real-LLM tests (require API key):
make llm-config # print resolved LLM config (no secrets)
make test-llm # end-to-end integration tests
make bench-appraisal # Scherer CPM prompt quality benchmarks
LLM test environment variables
| Variable | Default | Purpose |
|---|---|---|
EMOTIONAL_MEMORY_LLM_API_KEY |
— | API key (required) |
EMOTIONAL_MEMORY_LLM_BASE_URL |
https://api.openai.com/v1 |
OpenAI-compatible endpoint |
EMOTIONAL_MEMORY_LLM_MODEL |
gpt-5-mini |
Model |
EMOTIONAL_MEMORY_LLM_OUTPUT_MODE |
plain |
LLM response mode: plain or json_object |
EMOTIONAL_MEMORY_LLM_TIMEOUT_SECONDS |
30 |
HTTP timeout for OpenAI-compatible calls |
EMOTIONAL_MEMORY_LLM_REPEATS |
3 |
Repeats per phrase in quality benchmarks |
License
MIT — see LICENSE
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file emotional_memory-0.6.2.tar.gz.
File metadata
- Download URL: emotional_memory-0.6.2.tar.gz
- Upload date:
- Size: 131.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
05c3781b631153019131e6a52d5fe3d06e4cd81c72c942f3958ff5bc3b528691
|
|
| MD5 |
8ec26c2d339ebbfe18c7cdb1e9d7f021
|
|
| BLAKE2b-256 |
c534ab8305cf266f4b2bec6e5bd468d2334c6b403196a979eab28ccb3d3ec89d
|
File details
Details for the file emotional_memory-0.6.2-py3-none-any.whl.
File metadata
- Download URL: emotional_memory-0.6.2-py3-none-any.whl
- Upload date:
- Size: 80.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.13.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
19ce9649b2517e3e59fb8bf1e35347a7ee484082b0e1504116b9846507d6e51d
|
|
| MD5 |
3038f86e2597d346cac9c494fb71db54
|
|
| BLAKE2b-256 |
9595d877479c454a3392b59068850e953f934687bd1861dd724c5d1476c35f17
|