Skip to main content

Memory management library to enhance AI agents with smarter, personalized, context-aware responses

Project description

🧠 GPT Memory

Memory Driven Reasoning for Smarter AI Agents

GPT Memory is a library powered by Tovana that introduces a new approach to improving LLM reasoning through actionable insights (aka beliefs) derived from continuous interactions and long term memory.

PyPI version License: Apache 2

Why GPT Memory?

Current LLMs face significant limitations in their ability to learn and adapt from user-specific interactions over time. While LLMs excel at processing vast amounts of data, they struggle with ongoing personalization and context-aware learning. This gap restricts their ability to provide truly adaptive and evolving AI experiences.

Our Memory manager aims to address these challenges by providing a comprehensive memory and belief management framework for AI agents. Its core concept revolves around converting experiences (events) into memories, which in turn shape beliefs. These beliefs then influence the agent's reasoning, responses, and actions.

By simulating human-like memory processes, GPT Memory enables more personalized, adaptive, and context-aware AI interactions. This framework bridges the gap between static knowledge bases and dynamic, experience-based learning, allowing AI agents to evolve their understanding and behavior over time.

🌟 Features

Feature Status Description
🧠 Human-like Memory ✅ Available Transform interactions into lasting memories and actionable beliefs
🔍 Smart Information Extraction ✅ Available Automatically capture and store relevant user details from conversations
💡 Dynamic Belief Generation ✅ Available Create personalized, context-aware insights to guide AI responses
🤖 LLM-Friendly Context ✅ Available Seamlessly integrate memory and beliefs into your AI's decision-making process
🔌 Easy Integration ✅ Available Plug into your AI applications with a straightforward API
🎭 Conflict Resolution ✅ Available Intelligently handle contradictions in user information
🌐 Flexible Architecture ✅ Available Designed to work with various LLM providers and models
📊 Memory Management ✅ Available Process events, store short-term and long-term memories, and manage beliefs
🔗 Advanced Association Creation ✅ Available Form connections between memories and beliefs for more nuanced understanding
🧵 Async Functionality ✅ Available Support for asynchronous operations to enhance performance in concurrent environments
⛁ Persistent Database Support 🔜 Coming Soon Integration with persistent databases for long-term storage and retrieval of memory data
🎛️ Custom Belief Generation 🔜 Coming Soon User-generated beliefs offering end-to-end flexibility in shaping the belief system reasoning

🏗️ Architecture

Screenshot 2024-08-21 at 9 04 07

🚀 Quick Start

  1. Install Tovana:
pip install tovana
  1. Use it in your project:
from tovana import MemoryManager

business_description = "an AI therapist"
message = "I just moved from New York to Paris for work."
user_id = "user123"

# Initialize with your preferred LLM provider and API key (Refer to the documentation for specific models)
memory_manager = MemoryManager(api_key="your-llm-provider-api-key-here", provider="anthropic",
                               business_description=business_description, include_beliefs=True)

# Update user memory
memory_manager.update_user_memory(user_id=user_id, message=message)

# Get user memory
user_memory = memory_manager.get_user_memory(user_id=user_id)
print(user_memory)  # Output: {'location': 'Paris', 'previous_location': 'New York'}

# Get memory context for LLM
context = memory_manager.get_memory_context(user_id=user_id)
print(context)  # Output: 'User Memory:\n location: Paris,\n previous_location: New York'

# Get beliefs
beliefs = memory_manager.get_beliefs(user_id=user_id)
print(beliefs)  # Output: {"beliefs": "- Suggest spending time with Charlie and Luna when user is feeling down\n- Suggest family activities with Lisa and Mai for emotional well-being\n- Recommend playing basketball for physical exercise and stress relief"}

🧠 Belief Generation

GPT memory introduces a new approach to LLM reasoning: actionable beliefs generated from user memory. These beliefs provide personalized insights that can significantly enhance your agent's planning, reasoning and responses.

Examples

Input:

  • business_description: "a commerce site"
  • memory: {'pets': ['dog named charlie', 'horse named luna']}

Output:

{"beliefs": ",- suggest pet products for dogs and horses"}

Input:

  • business_description: "an AI therapist"
  • memory: {'pets': ['dog named charlie', 'horse named luna', 'sleep_time: 10pm']}

Output:

{"beliefs": ",- Suggest mediation at 9:30pm\n- Suggest spending time with Charlie and Luna for emotional well-being"}

🛠️ API Reference

MemoryManager

  • get_memory(user_id: str) -> JSON: Fetch user memory
  • delete_memory(user_id: str) -> bool: Delete user memory
  • update_memory(user_id: str, message: str) -> JSON: Update memory with relevant information if found in message
  • batch_update_memory(user_id: str, messages: List[Dict[str, str]]) -> JSON: Update memory with relevant information if found in message
  • get_memory_context(user_id: str, message: Optiona[str]) -> str: Get formatted memory context, general or message specific
  • get_beliefs(user_id: str) -> str: Get actionable beliefs context

Batch Update Memory

Traditional per-message memory updates can be costly and inefficient, especially in longer conversations. They often miss crucial context, leading to suboptimal information retrieval.

Our batch memory update method addresses these challenges by processing entire conversations at once. This approach not only improves performance and reduces costs but also enhances the quality of extracted information. This results in a more coherent and accurate user memory, ultimately leading to better AI reasoning.

Example

user_id = "user123"
messages = [
    {"role": "user", "content": "Hi, I'm planning a trip to Japan."},
    {"role": "assistant", "content": "That's exciting! When are you planning to go?"},
    {"role": "user", "content": "I'm thinking about next spring. I love sushi and technology."}
]

await memory_manager.batch_update_memory(user_id, messages)

Sync vs Async Updates

This library provides both synchronous and asynchronous update methods to cater to different use cases and application architectures:

  1. Asynchronous Updates (AsyncMemoryManager): Ideal for applications built on asynchronous frameworks like FastAPI or asynchronous Python scripts. This allows for non-blocking memory updates, improving overall application performance, especially when dealing with I/O-bound operations or high-concurrency scenarios.
  2. Synchronous Updates (MemoryManager): Suitable for traditional synchronous applications or when you need to ensure that memory updates are completed before proceeding with other operations. This can be useful in scripts or applications where the order of operations is critical.

By providing both options, our library offers flexibility, allowing to choose the most appropriate method based on your specific application requirements and architecture.

🤝 Contributing

We welcome contributions! Found a bug or have a feature idea? Open an issue or submit a pull request. Let's make Tovana even better together! 💪

📄 License

Tovana is Apache-2.0 licensed. See the LICENSE file for details.


Ready to empower your AI agents with memory-driven reasoning? Get started with GPT Memory! 🚀 If you find it useful, don't forget to star the repo! ⭐

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

tovana-0.0.8.tar.gz (18.1 kB view details)

Uploaded Source

Built Distribution

tovana-0.0.8-py3-none-any.whl (14.1 kB view details)

Uploaded Python 3

File details

Details for the file tovana-0.0.8.tar.gz.

File metadata

  • Download URL: tovana-0.0.8.tar.gz
  • Upload date:
  • Size: 18.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.12.5

File hashes

Hashes for tovana-0.0.8.tar.gz
Algorithm Hash digest
SHA256 f8358ba9d007343cc147f302784854064e72e5e69de2b564430d10091f9d2aa9
MD5 2dde76faff6935dc9b991db7b6080a22
BLAKE2b-256 35a8f0471b50fb6b2ffd88d927f0051170a18d2c859c61240cc2482e6b29537d

See more details on using hashes here.

File details

Details for the file tovana-0.0.8-py3-none-any.whl.

File metadata

  • Download URL: tovana-0.0.8-py3-none-any.whl
  • Upload date:
  • Size: 14.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.12.5

File hashes

Hashes for tovana-0.0.8-py3-none-any.whl
Algorithm Hash digest
SHA256 5c4ae48ea98f5aad9d905f6c595953f51e1ecc75fa7f9fb53edfba347cc4b53f
MD5 f755b7026b96bc4a5269395c98b8a306
BLAKE2b-256 40630fdf13665dbe13ec9d78ba5b045c1b902b8dd6721a1a6254255df41a2381

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page