Microsoft AutoGen integration for MCAL - Goal-aware memory for multi-agent systems
Project description
mcal-ai-autogen
Microsoft AutoGen integration for MCAL (Memory-Context Alignment Layer), bringing goal-aware memory to AutoGen agents.
Installation
pip install mcal-ai-autogen
# With AutoGen dependencies
pip install mcal-ai-autogen[autogen]
Quick Start
from autogen_agentchat.agents import AssistantAgent
from autogen_ext.models.openai import OpenAIChatCompletionClient
from mcal import MCAL
from mcal_autogen import MCALMemory
# Initialize MCAL
mcal = MCAL(llm_provider="openai")
# Create MCAL-backed memory
memory = MCALMemory(mcal, user_id="user_123")
# Create an agent with MCAL memory
model_client = OpenAIChatCompletionClient(model="gpt-4")
agent = AssistantAgent(
name="data_engineer",
model_client=model_client,
memory=[memory],
system_message="You are a helpful data engineering assistant.",
)
# Use the agent — MCAL automatically tracks context and decisions
result = await agent.run(task="How should I set up my ETL pipeline?")
What's New in 0.3.0
- Expanded Relationship Edge Types — 10 new edge types (
family,friend,colleague,likes,prefers,lives_in,works_at, etc.) for richer social and personal relationship graphs - Key Facts & Entities in Search Context —
search()now surfaces extracted facts and background entities directly inresult.context - Improved Chunk Retrieval — More results returned with equal weighting; conversation excerpts prioritized in context
- Higher Extraction Fidelity — Increased message processing window reduces information loss during ingestion
What's New in 0.2.9
- Configurable Extraction Profiles — Choose
decision(rationale/alternatives/trade-offs),conversational(preferences/relationships), orcomprehensive(both) - Hybrid Retrieval with ChunkStore — Graph traversal + embedding search for maximum recall
- FACT/PERSON Node Protection — Graph compaction preserves factual and identity nodes
- Benchmark Results — Decision profile achieves 93.3% detail retention, 62% token reduction in 150-turn CTO advisor tests
# New v0.2.8 — pass extraction options to MCAL
mcal = MCAL(
llm_provider="anthropic",
extraction_profile="decision",
enable_chunk_store=True,
)
memory = MCALMemory(mcal, user_id="user_123")
Features
Goal-Aware Memory
MCAL's unique value is understanding your project's goals and maintaining context across conversations:
mcal = MCAL(llm_provider="anthropic")
memory = MCALMemory(mcal)
# Add relevant context
from autogen_core.memory import MemoryContent
await memory.add(MemoryContent(
content="We decided to use Kafka for streaming",
mime_type="text/plain",
metadata={"category": "architecture", "decision": True}
))
# Query returns goal-relevant results
results = await memory.query("What messaging system should I use?")
# Returns Kafka decision with goal-relevance scoring
Decision Tracking
Track architectural and project decisions automatically:
memory = MCALMemory(
mcal,
enable_goal_tracking=True, # Extract goals from content
include_decisions=True, # Include decisions in search
)
# Decisions are automatically tracked
await memory.add(MemoryContent(
content="After evaluating options, we chose PostgreSQL for its JSON support",
mime_type="text/plain"
))
# Query finds relevant decisions
results = await memory.query("database selection")
User Isolation
Support multi-tenant scenarios with user isolation:
# Create separate memories for different users
user1_memory = MCALMemory(mcal, user_id="alice")
user2_memory = MCALMemory(mcal, user_id="bob")
# Each user has isolated memory
await user1_memory.add(MemoryContent(content="Alice prefers Python"))
await user2_memory.add(MemoryContent(content="Bob prefers Rust"))
# Queries only return user-specific results
results = await user1_memory.query("language preference")
# Only returns Alice's preference
TTL Support
Configure time-to-live for memory entries:
memory = MCALMemory(mcal, default_ttl_minutes=60) # 1 hour default
# Or per-entry TTL via metadata
await memory.add(MemoryContent(
content="Temporary context",
mime_type="text/plain",
metadata={"ttl_minutes": 15} # 15 minute TTL
))
Thread Safety
All operations are protected by RLock — safe for concurrent access from multiple agents.
Integration with AutoGen Features
With AssistantAgent
from autogen_agentchat.agents import AssistantAgent
agent = AssistantAgent(
name="assistant",
model_client=model_client,
memory=[memory], # MCAL memory integrates seamlessly
)
With Teams
from autogen_agentchat.teams import RoundRobinGroupChat
# Share MCAL memory across team members
shared_memory = MCALMemory(mcal, user_id="team_alpha")
coder = AssistantAgent("coder", model_client=model_client, memory=[shared_memory])
reviewer = AssistantAgent("reviewer", model_client=model_client, memory=[shared_memory])
team = RoundRobinGroupChat([coder, reviewer])
Context Window Management
MCAL automatically manages context relevance:
memory = MCALMemory(
mcal,
max_results=10, # Limit results per query
score_threshold=0.5, # Minimum relevance score
)
# update_context adds relevant memories to the agent's context
result = await memory.update_context(model_context)
API Reference
MCALMemory
class MCALMemory(Memory):
def __init__(
self,
mcal: MCAL,
user_id: str = "default",
name: str = "mcal_memory",
max_results: int = 10,
score_threshold: float = 0.0,
default_ttl_minutes: Optional[float] = None,
enable_goal_tracking: bool = True,
include_decisions: bool = True,
): ...
Key Methods
| Method | Async | Description |
|---|---|---|
add(content) |
✓ | Add MemoryContent to memory |
query(query) |
✓ | Search for relevant memories, returns MemoryQueryResult |
update_context(model_context) |
✓ | Update agent context with relevant memories |
clear() |
✓ | Clear all memory entries |
close() |
✓ | Cleanup resources |
Helper Methods
| Method | Description |
|---|---|
add_text(text, metadata=None) |
Convenience wrapper for adding plain text |
query_text(query) |
Convenience wrapper returning list of strings |
item_count |
Property returning number of stored items |
get_all_items() |
Return all non-expired memory items |
Requirements
- Python >= 3.11
- mcal-ai >= 0.2.0
- autogen-core >= 0.4.0 (optional — gracefully degrades if not installed)
- autogen-agentchat >= 0.4.0 (optional)
License
MIT License
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file mcal_ai_autogen-0.3.0.tar.gz.
File metadata
- Download URL: mcal_ai_autogen-0.3.0.tar.gz
- Upload date:
- Size: 14.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.11
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
fe2b154add2f142e1c4a47573e551f8bda0babd489223c5a14f078b96e3f7664
|
|
| MD5 |
41eadb8463ffff0e86c91f8e612d6346
|
|
| BLAKE2b-256 |
b31bad6e55f14c3cd07af2b92108c3dd3d182e403a1f7305ff2b78413be882a3
|
File details
Details for the file mcal_ai_autogen-0.3.0-py3-none-any.whl.
File metadata
- Download URL: mcal_ai_autogen-0.3.0-py3-none-any.whl
- Upload date:
- Size: 10.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.11
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
4b1747e551bf51a15c22d90ffd607543dbac545702c322172297d82a2e0d0628
|
|
| MD5 |
08ed066b0efcd17a1cc1f34f17671943
|
|
| BLAKE2b-256 |
ee7a3ab07cb8c72d0d158dbaa5fb0ebd07a5e0051a938318b8f61ed8dd7e6563
|