Skip to main content

Emotional memory library for LLMs based on Affective Field Theory

Project description

emotional_memory

CI PyPI Python DOI License: MIT

Emotional memory for LLMs based on Affective Field Theory (AFT) — a 5-layer model that encodes not just what happened, but how it felt, how that feeling was moving, and what mood colored the moment. Validated in English, Italian, and Spanish on the realistic-recall benchmark (N=200/80 per language).

Installation

uv pip install emotional-memory
uv pip install "emotional-memory[sentence-transformers]"  # real semantic embeddings (recommended)
uv pip install "emotional-memory[sqlite]"                 # SQLite persistence via sqlite-vec
uv pip install "emotional-memory[redis]"                  # shared affective-state persistence via Redis
uv pip install "emotional-memory[viz]"                    # matplotlib visualization
uv pip install "emotional-memory[dotenv]"                 # .env file loading via python-dotenv

For development:

git clone https://github.com/gianlucamazza/emotional-memory
cd emotional-memory
make install
# optional local demo stack:
make install-demo

Quickstart

uv pip install "emotional-memory[sentence-transformers]"
from emotional_memory import EmotionalMemory, InMemoryStore, CoreAffect
from emotional_memory.embedders import SentenceTransformerEmbedder

em = EmotionalMemory(
    store=InMemoryStore(),
    embedder=SentenceTransformerEmbedder(),  # all-MiniLM-L6-v2 by default
)

# Set current emotional state
em.set_affect(CoreAffect(valence=0.8, arousal=0.6))

# Encode memories — each one captures the full affective context
em.encode("Just shipped the feature after three hard weeks.")
em.encode("Team celebration in the office.", metadata={"source": "slack"})

# Retrieve — ranked by semantic relevance AND emotional congruence
results = em.retrieve("difficult project success", top_k=3)
for mem in results:
    print(mem.content, mem.tag.core_affect)

# Or inspect why a memory ranked where it did
explained = em.retrieve_with_explanations("difficult project success", top_k=1)
top = explained[0]
print(top.score)
print(top.breakdown.raw_signals)

Bring your own embedder — any object with .embed(text) -> list[float] works:

class MyEmbedder:
    def embed(self, text: str) -> list[float]: ...
    def embed_batch(self, texts: list[str]) -> list[list[float]]: ...

Or subclass SequentialEmbedder and implement only embed()embed_batch() is provided.

Async

import asyncio
from emotional_memory import EmotionalMemory, InMemoryStore, as_async
from emotional_memory.embedders import SentenceTransformerEmbedder

sync_em = EmotionalMemory(store=InMemoryStore(), embedder=SentenceTransformerEmbedder())
em = as_async(sync_em)  # wraps sync components with asyncio.to_thread bridges

async def main():
    await em.encode("Meeting went surprisingly well today.")
    results = await em.retrieve("work meeting", top_k=3)

asyncio.run(main())

For native async embedders or stores, construct AsyncEmotionalMemory directly with SyncToAsyncEmbedder, SyncToAsyncStore, or your own AsyncEmbedder/AsyncMemoryStore.

Affective Field Theory

AFT models emotion as a field — distributed, dynamic, multi-layer — rather than a discrete label or a single coordinate. Five layers are captured at encoding time:

Layer Model What it captures
CoreAffect Russell-Mehrabian PAD model Continuous (valence, arousal, dominance) — the emotional substrate
AffectiveMomentum Spinoza — affect as transition Velocity and acceleration of affect change
MoodField Heidegger — Stimmung as attunement Slow-moving global mood with inertia (EMA)
AppraisalVector Scherer/Lazarus/Stoics Emotion derived from evaluation: novelty, goal-relevance, coping, norm-congruence, self-relevance
ResonanceLinks Aristotle/Hume/Bower/Collins & Loftus/Hebb Associative bidirectional graph: semantic, emotional, temporal, causal, contrastive links; multi-hop spreading activation + Hebbian co-retrieval strengthening

Full theoretical foundations: docs/research/

API Overview

EmotionalMemory

em = EmotionalMemory(
    store: MemoryStore,
    embedder: Embedder,
    appraisal_engine: AppraisalEngine | None = None,  # optional: auto-appraise via LLM
    config: EmotionalMemoryConfig | None = None,
    state_store: AffectiveStateStore | None = None,   # optional: persist affective state
)
Method Description
encode(content, appraisal=None, metadata=None) -> Memory Encode content with full AFT pipeline
observe(content, appraisal=None, metadata=None) -> EmotionalTag Update affective state without storing a retrievable memory
encode_batch(contents, metadata=None) -> list[Memory] Batch encode with embed_batch(), per-item appraisal
retrieve(query, top_k=5) -> list[Memory] Emotionally-weighted retrieval + reconsolidation
retrieve_with_explanations(query, top_k=5) -> list[RetrievalExplanation] Same retrieval pipeline plus a structured score breakdown
elaborate(memory_id) -> Memory | None Run full appraisal on a fast-path (pending_appraisal=True) memory and blend core_affect
elaborate_pending() -> list[Memory] Elaborate all pending fast-path memories in one call
delete(memory_id) Remove a memory from the store
get(memory_id) -> Memory | None Look up a single memory by ID
list_all() -> list[Memory] Return all stored memories
len(engine) -> int Number of memories in the store
prune(threshold=0.05) -> int Delete memories below effective strength threshold; returns count removed
export_memories() -> list[dict] Serialise all memories to JSON-safe dicts (backup / migration)
import_memories(data, overwrite=False) -> int Restore from export_memories() output; returns count written
get_state() -> AffectiveState Current affective state (read-only copy)
set_affect(core_affect) Manually inject a CoreAffect
reset_state() Reset runtime affective state to the initial baseline
save_state() -> dict Serialise affective state for persistence
load_state(data) Restore previously saved affective state
persist_state() -> dict Persist the current affective state to the configured state store
restore_persisted_state() -> bool Restore the last persisted affective state from the configured state store
clear_persisted_state() Clear the configured persisted affective-state snapshot
get_current_mood(now=None) -> MoodField Read-only mood with time regression
close() Release store resources (e.g. SQLite connection); also via with

Both engines support context managers for automatic resource cleanup:

with EmotionalMemory(store=SQLiteStore("mem.db"), embedder=MyEmbedder()) as em:
    em.encode("Session start")
    results = em.retrieve("relevant context")
# SQLiteStore.close() called automatically

Use retrieve() for normal recall paths. Use retrieve_with_explanations() when you need the ranking-time decomposition (semantic, mood, affect, momentum, recency, resonance) for debugging, evaluation, or UI inspection.

AsyncEmotionalMemory

Same method signatures as EmotionalMemory. Coroutines: encode, observe, encode_batch, retrieve, retrieve_with_explanations, elaborate, elaborate_pending, delete, get, list_all, count, prune, export_memories, import_memories, persist_state, restore_persisted_state, clear_persisted_state, close. State accessors (get_state, set_affect, reset_state, save_state, load_state, get_current_mood) remain synchronous.

Supports async with for automatic resource cleanup:

async with AsyncEmotionalMemory(store=..., embedder=...) as em:
    await em.encode("Session start")
    results = await em.retrieve("relevant context")
from emotional_memory import AsyncEmotionalMemory, SyncToAsyncEmbedder, SyncToAsyncStore

Bridge adapters: SyncToAsyncEmbedder, SyncToAsyncStore, SyncToAsyncAppraisalEngine wrap any sync implementation. SyncToAsyncStore also proxies close() to the underlying store. as_async(engine) wraps a complete EmotionalMemory in one call.

Key config classes

  • EmotionalMemoryConfig — top-level config (decay, retrieval, resonance, mood_alpha, mood_decay)
  • RetrievalConfig — weights, APE threshold, reconsolidation learning rate
  • ResonanceConfig — similarity threshold, max links, semantic/emotional/temporal weights, candidate multiplier, propagation_hops, hebbian_increment, configurable link-classification thresholds
  • DecayConfig — power-law decay parameters, arousal modulation, floor values
  • MoodDecayConfig — time-based mood regression (half-life, inertia scale, baselines)
  • AdaptiveWeightsConfig — smooth mood-adaptive retrieval weight tuning (sigmoid/Gaussian gates)
  • LLMAppraisalConfig — LLM appraisal engine settings (system prompt, cache size, fallback behaviour)

Interfaces (bring your own)

If your embedder has no native batching, subclass SequentialEmbedderembed_batch() is provided automatically:

from emotional_memory import SequentialEmbedder

class MyEmbedder(SequentialEmbedder):
    def embed(self, text: str) -> list[float]:
        return my_model.encode(text).tolist()

Otherwise implement the full Embedder protocol:

class Embedder(Protocol):
    def embed(self, text: str) -> list[float]: ...
    def embed_batch(self, texts: list[str]) -> list[list[float]]: ...

class MemoryStore(Protocol):
    def save(self, memory: Memory) -> None: ...
    def get(self, memory_id: str) -> Memory | None: ...
    def update(self, memory: Memory) -> None: ...
    def delete(self, memory_id: str) -> None: ...
    def list_all(self) -> list[Memory]: ...
    def search_by_embedding(self, embedding: list[float], top_k: int) -> list[Memory]: ...
    def __len__(self) -> int: ...

class AffectiveStateStore(Protocol):
    def save(self, state: AffectiveState) -> None: ...
    def load(self) -> AffectiveState | None: ...
    def clear(self) -> None: ...

Async variants (AsyncEmbedder, AsyncMemoryStore, AsyncAppraisalEngine) are defined in interfaces_async.py. AsyncMemoryStore uses count() -> int instead of __len__ since dunder methods cannot be coroutines.

Stores included:

  • InMemoryStore — dict-backed, brute-force cosine search (no extra deps)
  • SQLiteStore — persistent SQLite + sqlite-vec ANN search (uv pip install "emotional-memory[sqlite]")

Appraisal Engines

class AppraisalEngine(Protocol):
    def appraise(self, event_text: str, context: dict | None = None) -> AppraisalVector: ...

Pass an appraisal_engine to EmotionalMemory to auto-generate AppraisalVector during encode.

LLMAppraisalEngine — wrap any LLM SDK in a single callable:

from emotional_memory import LLMAppraisalEngine

def my_llm(prompt: str, json_schema: dict) -> str:
    # call openai / anthropic / local model here
    return response_text

engine = LLMAppraisalEngine(llm=my_llm)
em = EmotionalMemory(store=..., embedder=..., appraisal_engine=engine)

KeywordAppraisalEngine — regex-based fallback, zero external dependencies, ships with default rules covering success, failure, novelty, danger, and social norms:

from emotional_memory import KeywordAppraisalEngine
engine = KeywordAppraisalEngine()  # or pass custom KeywordRule list

Visualization

The optional viz extra provides 8 plotting functions for inspecting and presenting the model's internals. Each function accepts an optional ax parameter for subplot composition and returns a matplotlib.Figure.

from emotional_memory.visualization import plot_circumplex, plot_decay_curves

Valence-Arousal Circumplex

Memories plotted on the Russell-Mehrabian PAD model (valence-arousal plane), colored by consolidation strength.

Circumplex

Decay Curves (ACT-R Power Law)

Family of curves showing how arousal (McGaugh 2004) and retrieval count (spacing effect) modulate memory decay.

Decay Curves

Yerkes-Dodson Inverted-U

Consolidation strength peaks near effective arousal 0.7, then drops — the classic Yerkes-Dodson curve.

Yerkes-Dodson

6-Signal Retrieval Breakdown

Radar chart of the six retrieval signals: semantic similarity, mood congruence, affect proximity, momentum alignment, recency, and resonance boost.

Retrieval Radar

Mood Field Evolution

Time series of valence, arousal, and dominance with dashed baselines showing the regression attractors.

Mood Evolution

Adaptive Retrieval Weights

Heatmap showing how retrieval weights shift across different mood states (valence x arousal grid).

Adaptive Weights

Resonance Network

Directed graph with memories as nodes and edges colored by link type (semantic, emotional, temporal, causal, contrastive).

Resonance Network

Appraisal Radar (Scherer CPM)

Spider chart of the 5 Stimulus Evaluation Check dimensions.

Appraisal Radar

Generating images

make docs-images   # regenerate all PNGs in docs/images/
make research-figures   # regenerate benchmark evidence figures

Evidence Figures

The research figures below are generated from committed benchmark JSON artefacts, not from rerunning long studies.

Realistic Replay Overview

Realistic Replay Challenge Breakdown

S3 Ablation Study

Italian Multilingual Slice

LoCoMo Negative Result

Comparison with Existing Systems

emotional-memory (AFT) Mem0 Letta Zep LangChain Memory
License MIT Apache 2.0 Apache 2.0 Apache 2.0 MIT
Persistence InMemory / SQLite Qdrant, Chroma, Pinecone, PG, MongoDB PostgreSQL / SQLite Neo4j (self-hosted) / Cloud In-memory / custom
BYO embedder ✅ any Embedder protocol ✅ (OpenAI default) ⚠️ partial ⚠️ partial
Emotion model ✅ 5-layer AFT (valence, arousal, dominance, mood, appraisal, resonance)
Reconsolidation ✅ APE-gated lability window ✅ auto update/remove ✅ tool-call edit ✅ edge invalidation
Persistent mood state ✅ MoodField (Heidegger EMA)
LLM-agnostic
LangChain integration EmotionalMemoryChatHistory ✅ official ✅ tools interop ✅ ZepVectorStore ✅ native
Internal fidelity tests ✅ 126 cases, 20 phenomena (bench-fidelity)
External benchmark ✅ LoCoMo (FAIL: F1 0.168 vs 0.271) ✅ LoCoMo, LongMemEval, BEAM ✅ LoCoMo, DMR ✅ DMR, LongMemEval
Codebase size ~4.8k LOC (src/) >50k LOC >50k LOC >50k LOC ~5k LOC

Key differentiator: emotional-memory makes affect a first-class, multi-layer part of encoding and retrieval. Compared with the general-purpose memory systems in this table, it emphasizes mood-congruent retrieval, appraisal-conditioned tagging, and APE-gated reconsolidation rather than generic conversational recall alone.

Benchmark caveat: AFT fidelity tests validate psychological invariants (intra-theory). The comparative benchmark in this repo is a controlled synthetic retrieval probe, not a general downstream evaluation of production memory systems.

Current validation status

  • Strongest evidence today: 126 fidelity test cases across 20 phenomena show that the implementation behaves coherently with the theories it operationalizes.
  • Comparative evidence today: the repo includes a synthetic affect-aware retrieval benchmark (affect_reference_v1) and a realistic multi-session replay benchmark (v2, 50 scenarios, 200 queries, 5 challenge types × 40). On the v2 benchmark, AFT outperforms naive_cosine on both embedder classes: SBERT bge-small-en-v1.5 — top1 0.53 vs 0.33, Δ=+0.21 [0.15,0.27], p<0.001, d=0.49; e5-small-v2 — top1 0.50 vs 0.34, Δ=+0.16 [0.09,0.22], p<0.001, d=0.31. Architecture attribution is confirmed (appraisal confound ruled out, Gate 3 CLOSED). On the controlled quadrant probe (affect_reference_v1, SBERT embedder), AFT and naive_cosine both reach recall@5 = 0.80 (ceiling effect at N = 20 items; both well above recency baseline 0.25).
  • Italian multilingual slice (G6, 20 scenarios / 80 queries): With SBERT bge-small-en-v1.5 (EN-only): AFT top1=0.24 [0.14,0.33], hit@k=0.34 [0.24,0.44]; naive_cosine top1=0.15, hit@k=0.19. Δ hit@k=+0.15 [0.07,0.24], p=0.0005 (signal holds); top1 Δ not significant. With multilingual-e5-small (intfloat, 117M params, 100+ languages): AFT top1=0.29 [0.20,0.39], hit@k=0.42 [0.31,0.54]; naive_cosine top1=0.21, hit@k=0.26. Δ hit@k=+0.16 [0.06,0.26], p=0.001 (signal preserved); top1 Δ not significant. The embedder swap lifts naive_cosine top1 from 0.15→0.21 (+40%), confirming the absolute-accuracy gap was driven by the EN-only embedder, not AFT. Residual gap vs English v2 (top1 0.53) reflects dataset difficulty and model size; a larger multilingual model (e.g. bge-m3) is a natural next step.
  • Spanish multilingual slice (Hd2_ES, 20 scenarios / 80 queries): With SBERT bge-small-en-v1.5: AFT top1=0.25, naive_cosine top1=0.10, Δ=+0.138 [p=0.045, d=0.233] — PASS. Hypothesis Hd2 generalizes to Spanish with SBERT. With multilingual-e5-small: Δ=+0.113 [p=0.110, d=0.189] — FAIL (marginal, borderline signal; larger multilingual model expected to close gap).
  • Negative external result (LoCoMo, Gate 1 FAIL): on the LoCoMo conversational QA benchmark (1986 QA pairs, 10 conversations), AFT underperforms a naive RAG baseline (F1 0.168 vs 0.271). Affective weighting does not help on factual open-domain QA.
  • Not yet established: general superiority over systems such as Mem0 or LangMem on realistic agent tasks; completed human evaluation results; or ecological correspondence to human emotional memory.

See Current Evidence for the study ladder and the current claim-to-evidence matrix. The canonical machine-readable source for those public scientific claims lives in docs/research/claim_validation_matrix.json.

Benchmarks

Psychological fidelity (126 parametrized test cases, 20 phenomena)

The library validates 20 phenomena from the affective science literature via 126 parametrized test cases (run pytest --collect-only benchmarks/fidelity/ to enumerate them):

Phenomenon Reference Cases Test file
Mood-congruent recall Bower 1981 3 test_mood_congruent.py
Emotional enhancement Cahill & McGaugh 1995 3 test_emotional_enhancement.py
Yerkes-Dodson inverted-U Yerkes & Dodson 1908 12 test_yerkes_dodson.py
Spacing effect Ebbinghaus 1885 7 test_spacing_effect.py
Arousal floor McGaugh 2004 7 test_arousal_floor.py
Reconsolidation (APE) Nader & Schiller 2000 5 test_reconsolidation.py
State-dependent retrieval Godden & Baddeley 1975 3 test_state_dependent.py
Affective momentum Spinoza, Ethics III 9 test_momentum.py
Mood-adaptive weights Heidegger, Being & Time §29 14 test_mood_adaptive.py
Appraisal-to-affect mapping Scherer CPM 2009 11 test_appraisal_affect.py
Spreading activation Collins & Loftus 1975 5 test_spreading_activation.py
Hebbian co-retrieval strengthening Hebb 1949 4 test_hebbian_strengthening.py
ACT-R power-law decay Anderson 1983 / McGaugh 2004 5 test_decay_power_law.py
PAD dominance Mehrabian & Russell 1974 8 test_pad_dominance.py
Emotional retrieval vs. cosine Bower 1981 / Russell 1980 / Nader 2000 3 test_emotional_vs_cosine.py
Design gap regression (various) 3 test_design_gaps.py
Dual-path encoding LeDoux 1996 6 test_dual_path_encoding.py
Emotion categorization Plutchik 1980 10 test_emotion_categorization.py
Affective prediction error Schultz 1997 / Pearce-Hall 1980 5 test_prediction_error.py
APE-gated reconsolidation window Nader & Schiller 2000 3 test_reconsolidation_window.py

Run with: make bench-fidelity

For the comparative protocol and interpretation rules, see benchmarks/comparative/protocol.md.

Performance (hash-based embedder, InMemoryStore)

Operation N Mean OPS
Encode (single) 1 1.7 ms 590/s
Encode (batch of 100) 100 9.9 ms/op 101/s
Encode w/ resonance graph 500 4.0 ms 250/s
Retrieve top-5 100 ~2 ms ~500/s
Retrieve top-5 1 000 ~12 ms ~85/s
Retrieve top-5 10 000 ~120 ms ~8/s
Retrieve (top-k 1–25) 1 000 10–18 ms 55–100/s
Retrieve + reconsolidation 200 2.6 ms 385/s

InMemoryStore.search_by_embedding uses vectorized matrix multiplication (numpy), making retrieval O(n · d) in a single batch rather than n individual cosine calls. Retrieval uses two-pass scoring (spreading activation); when no resonance links are active the second pass is skipped. For stores > 10 000 memories, use SQLiteStore (sqlite-vec ANN) or a vector database implementing the MemoryStore protocol.

Run with: make bench-perf

Appraisal quality (LLM prompt validation)

15 natural-language phrases with expected directional outcomes against Scherer's 5 dimensions:

Phrase category Key assertions
Personal loss ("I got fired") goal_relevance < -0.2, coping_potential < 0.6
Achievement ("Got promoted") goal_relevance > 0.2, norm_congruence > 0.0
Moral violation ("Coworker stole credit") norm_congruence < -0.2, goal_relevance < 0.0
Grief, danger, betrayal, relief, … dimension-specific directional bounds

Assertions use wide bands (e.g. > 0.3, < -0.2) and evaluate the median over 3 LLM calls to tolerate non-determinism. Designed to catch systematic prompt regressions, not exact calibration.

Run with: EMOTIONAL_MEMORY_LLM_API_KEY=... make bench-appraisal

Works with any OpenAI-compatible endpoint (Ollama, vLLM, LiteLLM, …) via EMOTIONAL_MEMORY_LLM_BASE_URL.

LangChain integration

EmotionalMemoryChatHistory is a drop-in replacement for any LangChain chat history object. It backs the transcript with an EmotionalMemory instance so the affective state evolves naturally as the conversation unfolds, while letting you control which messages become retrievable memories.

uv pip install "emotional-memory[langchain,sentence-transformers]"
from emotional_memory import EmotionalMemory, InMemoryStore
from emotional_memory.embedders import SentenceTransformerEmbedder
from emotional_memory.integrations import (
    EmotionalMemoryChatHistory,
    recommended_conversation_policy,
)

em = EmotionalMemory(
    store=InMemoryStore(),
    embedder=SentenceTransformerEmbedder(),
)
history = EmotionalMemoryChatHistory(em, message_policy=recommended_conversation_policy)

# Works anywhere BaseChatMessageHistory is accepted:
history.add_user_message("I'm anxious about the deadline.")
history.add_ai_message("Let's break the work into smaller steps.")

print(history.messages)   # [HumanMessage(...), AIMessage(...)]

# The underlying engine has tracked the affective state:
state = em.get_state()
print(f"valence={state.core_affect.valence:.2f}  arousal={state.core_affect.arousal:.2f}")

With recommended_conversation_policy, user messages become retrievable memories, assistant messages update affective state without being stored, and control commands such as recall ... are ignored by retrieval. The adapter uses dependency injection — pass a fully-configured EmotionalMemory so you control the store backend and embedder. clear() removes stored memories, clears the transcript, and resets affective state.

Logging

The library uses the standard logging module. Enable debug output to trace the full pipeline:

import logging
logging.basicConfig(level=logging.DEBUG)
# or just for emotional_memory:
logging.getLogger("emotional_memory").setLevel(logging.DEBUG)

Debug events include: encode start/stored/resonance, retrieve start/done, reconsolidation triggers, LLM appraisal cache hits, and fallback activations.

Examples

The examples/ directory contains runnable scripts covering the full API. All scripts are self-contained and use a deterministic HashEmbedder so they run without any ML dependencies.

Script Description Extra
basic_usage.py Encode/retrieve, reconsolidation, resonance links
advanced_config.py ACT-R decay, mood regression, adaptive weights
appraisal_engines.py Keyword, static, and custom appraisal rules
reconsolidation.py Two-retrieval lability window (Nader & Schiller 2000)
async_usage.py as_async(), SyncToAsync* adapters, encode_batch
persistence.py SQLiteStore, save/load state, export/import, prune [sqlite]
emotional_journal.py Multi-session journaling with full lifecycle [sqlite]
llm_appraisal.py LLM-backed appraisal via OpenAI-compatible API openai
httpx_llm_integration.py httpx LLMCallable, .env config, 7 API deep-dives httpx
sentence_transformers_embedder.py SequentialEmbedder with real embeddings sentence-transformers
visualization.py All 8 matplotlib plot types [viz]
resonance_network.py Resonance graph and link-type distribution [viz]
retrieval_signals.py 6-signal decomposition, radar chart, weight heatmap [viz]

Run any script: uv run python examples/<script>.py

Development

make check                    # lint + typecheck + test
make cov                      # tests with branch coverage report
make bench                    # fidelity + performance benchmarks

# Real-LLM tests (require API key):
make llm-config                # print resolved LLM config (no secrets)
make test-llm                 # end-to-end integration tests
make bench-appraisal          # Scherer CPM prompt quality benchmarks

LLM test environment variables

Variable Default Purpose
EMOTIONAL_MEMORY_LLM_API_KEY API key (required)
EMOTIONAL_MEMORY_LLM_BASE_URL https://api.openai.com/v1 OpenAI-compatible endpoint
EMOTIONAL_MEMORY_LLM_MODEL gpt-5-mini Model
EMOTIONAL_MEMORY_LLM_REASONING_EFFORT "" Reasoning budget for o-series / gpt-5 models (minimal / low / medium / high); omitted when empty
EMOTIONAL_MEMORY_LLM_OUTPUT_MODE plain LLM response mode: plain or json_object
EMOTIONAL_MEMORY_LLM_TIMEOUT_SECONDS 30 HTTP timeout for OpenAI-compatible calls
EMOTIONAL_MEMORY_LLM_REPEATS 3 Repeats per phrase in quality benchmarks

License

MIT — see LICENSE

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

emotional_memory-0.8.3.tar.gz (159.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

emotional_memory-0.8.3-py3-none-any.whl (88.1 kB view details)

Uploaded Python 3

File details

Details for the file emotional_memory-0.8.3.tar.gz.

File metadata

  • Download URL: emotional_memory-0.8.3.tar.gz
  • Upload date:
  • Size: 159.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.9.15 {"installer":{"name":"uv","version":"0.9.15","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Arch Linux","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for emotional_memory-0.8.3.tar.gz
Algorithm Hash digest
SHA256 7e866aa0e821b42caecbddbfbfc2636ba7f3c9500a699190250a19033730d3fb
MD5 b2318e110f99abedd822c8a40ea2248c
BLAKE2b-256 8dd742ad0fe1b98e076e4066bdba5960eaf04a89994d82c436c0e098db4579ce

See more details on using hashes here.

File details

Details for the file emotional_memory-0.8.3-py3-none-any.whl.

File metadata

  • Download URL: emotional_memory-0.8.3-py3-none-any.whl
  • Upload date:
  • Size: 88.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.9.15 {"installer":{"name":"uv","version":"0.9.15","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Arch Linux","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for emotional_memory-0.8.3-py3-none-any.whl
Algorithm Hash digest
SHA256 e766a241566d7cc8290c3d223a3d2cc5685552c4935c0565a5c7302682094753
MD5 294355272c7252fd21eb936ac50ad44d
BLAKE2b-256 00b9c5b3427aa2a031499470df9087632f81fa8021c0dbd04e65e2387dae92a6

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page