Skip to main content

Session manager for AI applications

Project description

CHUK AI Session Manager

A powerful session management system for AI applications

Python 3.11+ License: Apache 2.0

Automatic conversation tracking, token usage monitoring, tool call logging, infinite context support with automatic summarization, and hierarchical session relationships. Perfect for AI applications that need reliable session management.

🚀 Quick Start

Installation Options

# Basic installation (memory storage only)
pip install chuk-ai-session-manager

# With Redis support for persistent storage
pip install chuk-ai-session-manager[redis]

# With enhanced token counting
pip install chuk-ai-session-manager[tiktoken]

# Full installation with all optional features
pip install chuk-ai-session-manager[all]

# Development installation
pip install chuk-ai-session-manager[dev]

Quick Example

from chuk_ai_session_manager import track_conversation
from chuk_ai_session_manager.config import DEFAULT_TOKEN_MODEL

# Track any conversation automatically
session_id = await track_conversation(
    user_message="What's the weather like?",
    ai_response="I don't have access to real-time weather data.",
    model=DEFAULT_TOKEN_MODEL,
    provider="openai"
)

print(f"Conversation tracked in session: {session_id}")

That's it! Zero configuration required.

⚡ Major Features

🧠 AI Virtual Memory

OS-style memory management for AI context windows. Pages, working sets, faults, eviction, and compression — giving conversations the illusion of infinite memory.

from chuk_ai_session_manager import SessionManager
from chuk_ai_session_manager.config import DEFAULT_TOKEN_MODEL
from chuk_ai_session_manager.memory import (
    MemoryManager, CompressorRegistry, ImportanceWeightedLRU,
    PageType, VMMode, WorkingSetConfig,
)

# Zero-config: just enable VM on SessionManager
sm = SessionManager(enable_vm=True, vm_mode=VMMode.STRICT)

# Or fully customize eviction and compression
vm = MemoryManager(
    session_id="my_session",
    config=WorkingSetConfig(max_l0_tokens=32_000),
    eviction_policy=ImportanceWeightedLRU(),       # Swappable strategy
    compressor_registry=CompressorRegistry.default(), # Per-modality compression
)

# Create pages, add to working set
page = vm.create_page("Decision: Use JWT for auth", page_type=PageType.CLAIM)
await vm.add_to_working_set(page)

# Build context for LLM call
ctx = vm.build_context(system_prompt="You are helpful.")
# ctx["developer_message"] has VM:RULES + VM:MANIFEST_JSON + VM:CONTEXT

Eviction policies: ImportanceWeightedLRU (default), LRUEvictionPolicy, ModalityAwareLRU — or implement the EvictionPolicy protocol for custom strategies.

Compression: Pages compress through FULL → REDUCED → ABSTRACT → REFERENCE before eviction, saving tokens without losing context. Text, image, and passthrough compressors included; plug in custom summarize_fn for LLM-based compression.

See AI Virtual Memory docs for full documentation.

🎯 Zero-Configuration Tracking

from chuk_ai_session_manager import SessionManager
from chuk_ai_session_manager.config import DEFAULT_TOKEN_MODEL

# Just start using it
sm = SessionManager()
await sm.user_says("Hello!")
await sm.ai_responds("Hi there!", model=DEFAULT_TOKEN_MODEL)

# Get stats instantly
stats = await sm.get_stats()
print(f"Tokens: {stats['total_tokens']}, Cost: ${stats['estimated_cost']:.4f}")

🔄 Infinite Context

# Automatically handles conversations longer than token limits
sm = SessionManager(infinite_context=True, token_threshold=4000)
await sm.user_says("Tell me about the history of computing...")
await sm.ai_responds("Computing history begins with...", model=DEFAULT_TOKEN_MODEL)
# Session will auto-segment when limits are reached

⚙️ Storage Backends

Installation Storage Use Case Performance
pip install chuk-ai-session-manager Memory Development, testing 1.8M ops/sec
pip install chuk-ai-session-manager[redis] Redis Persistent, distributed 20K ops/sec

🛡️ Conversation Guards and Tool State

Runtime guardrails that prevent runaway tool loops, track value bindings, and enforce grounded tool calls.

from chuk_ai_session_manager.guards import get_tool_state, ToolStateManager

# Get the singleton tool state manager
tool_state = get_tool_state()

# Track tool calls and bind results as $v1, $v2, ...
binding = tool_state.bind_value("sqrt", {"x": 16}, 4.0)
# LLM can now reference $v1 in subsequent calls

# Check for runaway tool loops
status = tool_state.check_runaway()

# Detect ungrounded calls (missing $vN references)
check = tool_state.check_ungrounded_call("normal_cdf", {"mean": 0, "std": 1, "x": 1.5})

# Reset state for a new prompt
tool_state.reset_for_new_prompt()

Guard components:

  • ToolStateManager - Coordinator for all guards, bindings, and cache
  • BindingManager - $vN reference system for tracking tool results
  • ResultCache - Tool result caching for deduplication
  • UngroundedGuard - Detects calls with missing computed-value references
  • Runtime guards (budget, runaway, per-tool limits) from chuk-tool-processor

🧩 Procedural Memory

Learn from tool call history to improve future tool use.

from chuk_ai_session_manager import ToolMemoryManager, ProceduralContextFormatter

# Record tool outcomes
memory = ToolMemoryManager()
await memory.record("calculator", {"op": "add", "a": 5, "b": 3}, result=8, success=True)

# Format learned patterns for the model's context
formatter = ProceduralContextFormatter()
context = formatter.format(memory.get_patterns())

🛠️ Tool Integration

# Automatic tool call tracking
await sm.tool_used(
    tool_name="calculator",
    arguments={"operation": "add", "a": 5, "b": 3},
    result={"result": 8}
)

💡 Common Use Cases

Web App Conversation Tracking

from chuk_ai_session_manager import track_conversation
from chuk_ai_session_manager.config import DEFAULT_TOKEN_MODEL

# In your chat endpoint
session_id = await track_conversation(
    user_message=request.message,
    ai_response=ai_response,
    model=DEFAULT_TOKEN_MODEL,
    provider="openai",
    session_id=request.session_id  # Continue existing conversation
)

LLM Wrapper with Automatic Tracking

from chuk_ai_session_manager import track_llm_call
from chuk_ai_session_manager.config import DEFAULT_TOKEN_MODEL
import openai

async def my_openai_call(prompt):
    response = await openai.chat.completions.create(
        model=DEFAULT_TOKEN_MODEL,
        messages=[{"role": "user", "content": prompt}]
    )
    return response.choices[0].message.content

# Automatically tracked
response, session_id = await track_llm_call(
    user_input="Explain machine learning",
    llm_function=my_openai_call,
    model=DEFAULT_TOKEN_MODEL,
    provider="openai"
)

Long Conversations with Auto-Segmentation

from chuk_ai_session_manager import track_infinite_conversation
from chuk_ai_session_manager.config import DEFAULT_TOKEN_MODEL # gpt-4o, gpt-4o-mini, etc.

# Start a conversation
session_id = await track_infinite_conversation(
    user_message="Tell me about the history of computing",
    ai_response="Computing history begins with ancient calculating devices...",
    model=DEFAULT_TOKEN_MODEL,
    token_threshold=4000  # Auto-segment after 4000 tokens
)

# Continue the conversation - will auto-segment if needed
session_id = await track_infinite_conversation(
    user_message="What about quantum computers?",
    ai_response="Quantum computing represents a fundamental shift...",
    session_id=session_id,
    model=DEFAULT_TOKEN_MODEL
)

🔧 Configuration

Storage Configuration

# Memory provider (default) - fast, no persistence
export SESSION_PROVIDER=memory

# Redis provider - persistent, distributed (requires redis extra)
export SESSION_PROVIDER=redis
export SESSION_REDIS_URL=redis://localhost:6379/0

Model Configuration

DEFAULT_TOKEN_MODEL resolves to the default LLM model defined in:

  • config.py
  • or the CHUK_DEFAULT_MODEL environment variable

By default (as shipped), it is set to:

  • gpt-4o-mini

You can override it explicitly:

await sm.ai_responds(
    "Hello!",
    model="gpt-4o"
)

The selected model must be compatible with the specified provider.

Installation Matrix

Command Memory Redis Token Counting Use Case
pip install chuk-ai-session-manager Basic Development
pip install chuk-ai-session-manager[redis] Basic Persistent
pip install chuk-ai-session-manager[tiktoken] Enhanced Better accuracy
pip install chuk-ai-session-manager[all] Enhanced Full features

📊 Monitoring & Analytics

# Get comprehensive session analytics
stats = await sm.get_stats(include_all_segments=True)

print(f"""
🚀 Session Analytics Dashboard
============================
Session ID: {stats['session_id']}
Total Messages: {stats['total_messages']}
User Messages: {stats['user_messages']}
AI Messages: {stats['ai_messages']}
Tool Calls: {stats['tool_calls']}
Total Tokens: {stats['total_tokens']}
Total Cost: ${stats['estimated_cost']:.6f}
Session Segments: {stats.get('session_segments', 1)}
""")

🏗️ Why CHUK AI Session Manager?

  • Zero Configuration: Start tracking conversations in 3 lines of code
  • Infinite Context: Never worry about token limits again
  • Universal: Works with any LLM provider (OpenAI, Anthropic, etc.)
  • Robust: Built-in persistence, monitoring, and error handling
  • Token Aware: Automatic cost tracking across all providers
  • Tool Friendly: Seamless tool call logging and retry mechanisms
  • Guardrails: Runtime guards prevent runaway tool loops and ungrounded calls
  • Procedural Memory: Learn from tool call history to improve future use

🛡️ Error Handling

from chuk_ai_session_manager import (
    SessionManagerError,
    SessionNotFound,
    TokenLimitExceeded
)

try:
    session_id = await track_conversation("Hello", "Hi there")
except SessionNotFound as e:
    print(f"Session not found: {e}")
except TokenLimitExceeded as e:
    print(f"Token limit exceeded: {e}")
except SessionManagerError as e:
    print(f"General session error: {e}")

🔄 Dependencies

  • Required: chuk-sessions (session storage), pydantic (data models), chuk-tool-processor (tool integration)
  • Optional: redis (Redis storage), tiktoken (accurate token counting)

📄 License

Apache 2.0 - build amazing AI applications with confidence!


Ready to build better AI applications?

pip install chuk-ai-session-manager

Start tracking conversations in 30 seconds!

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

chuk_ai_session_manager-0.11.1.tar.gz (239.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

chuk_ai_session_manager-0.11.1-py3-none-any.whl (150.3 kB view details)

Uploaded Python 3

File details

Details for the file chuk_ai_session_manager-0.11.1.tar.gz.

File metadata

  • Download URL: chuk_ai_session_manager-0.11.1.tar.gz
  • Upload date:
  • Size: 239.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.11.11

File hashes

Hashes for chuk_ai_session_manager-0.11.1.tar.gz
Algorithm Hash digest
SHA256 702aedbf01c1c9188eedc4d579794bc1935271f8c35a7afa9a97ac6ff6161ae9
MD5 d4c5dc95b5b7def82c115110cf6b1c17
BLAKE2b-256 1f3ff73eb0f2f5ee236323a7bae92abd004ba7f7775621c5a84e4d900fbbc291

See more details on using hashes here.

File details

Details for the file chuk_ai_session_manager-0.11.1-py3-none-any.whl.

File metadata

File hashes

Hashes for chuk_ai_session_manager-0.11.1-py3-none-any.whl
Algorithm Hash digest
SHA256 8264459d8672263cc21e5a439d7f92417b687ff6f11aeecdbf0b2f579f080e6e
MD5 501ed96ed488af173f4b99c37332f858
BLAKE2b-256 49e2856047cf98de1a0cf2553d14d48fff34556607bbc4af5195a1a71ea8e185

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page