A smart conversation memory management system
Project description
MFCS Memory
MFCS Memory is an intelligent conversation memory management system that helps AI assistants remember conversation history with users and dynamically adjust response strategies based on conversation content.
Key Features
- Intelligent Conversation Memory: Automatically analyzes and summarizes user characteristics and preferences
- Vector Storage: Uses Qdrant for efficient similar conversation retrieval
- Session Management: Supports multi-user, multi-session management
- Automatic Chunking: Automatically creates chunks when conversation history exceeds threshold
- Async Support: All operations support asynchronous execution
- Extensibility: Modular design, easy to extend and customize
- Automatic LLM-based Analysis: User memory and conversation summary are updated automatically at configurable intervals
Core Modules
core/base.py: Base manager, handles all shared connections (MongoDB, Qdrant, embedding model)core/conversation_analyzer.py: Analyzes conversation content and user profile using LLM (OpenAI API)core/memory_manager.py: Main entry for memory management, orchestrates all modules and async taskscore/session_manager.py: Handles session creation, update, chunking, and analysis task managementcore/vector_store.py: Handles vector storage, retrieval, and chunked dialog managementutils/config.py: Loads and validates all configuration from environment variables
Core Features
MemoryManager Core Methods
-
get(user_id: str, query: Optional[str] = None, top_k: int = 2) -> str
- Get current session information for specified user
- Includes conversation summary and user memory summary
- Supports query-based relevant historical conversation retrieval (vector search)
- Returns formatted memory information
-
update(user_id: str, user_input: str, assistant_response: str) -> bool
- Automatically gets or creates current session for user
- Updates conversation history
- Automatically updates user memory summary every 3 rounds (LLM analysis)
- Automatically updates session summary every 5 rounds (LLM analysis)
- Automatically handles conversation chunking and vector storage
- All analysis tasks run asynchronously and are recoverable on restart
-
delete(user_id: str) -> bool
- Deletes all data for specified user (session + vector store)
- Returns whether operation was successful
-
reset() -> bool
- Resets all user records (clears all session and vector data)
- Returns whether operation was successful
Installation
- Install the package:
pip install mfcs-memory
- Install SentenceTransformer for text embedding:
pip install sentence-transformers
Note: The default embedding model is
BAAI/bge-large-zh-v1.5. You can change it in the configuration.
Quick Start
- Create a
.envfile and configure necessary environment variables:
# MongoDB Configuration
MONGO_USER=your_username
MONGO_PASSWD=your_password
MONGO_HOST=localhost:27017
# Qdrant Configuration
QDRANT_HOST=localhost
QDRANT_PORT=6333
# Model Configuration
EMBEDDING_MODEL_PATH=./model/BAAI/bge-large-zh-v1.5
EMBEDDING_DIM=768
LLM_MODEL=qwen-plus-latest # Default value
# OpenAI Configuration
OPENAI_API_KEY=your_api_key
OPENAI_API_BASE=your_api_base # Optional
# Other Configuration
MONGO_REPLSET='' # Optional, if using replica set
MAX_RECENT_HISTORY=20 # Default value
CHUNK_SIZE=100 # Default value
MAX_CONCURRENT_ANALYSIS=3 # Default value
- Usage Example:
import asyncio
from mfcs_memory.utils.config import Config
from mfcs_memory.core.memory_manager import MemoryManager
async def main():
# Load configuration
config = Config.from_env()
# Initialize memory manager
memory_manager = MemoryManager(config)
# Update conversation
await memory_manager.update(
"user123",
"Hello, I want to learn about Python programming",
"Python is a simple yet powerful programming language..."
)
# Get memory information
memory_info = await memory_manager.get(
"user123",
query="How to start Python programming?",
top_k=2
)
# Delete user data
await memory_manager.delete("user123")
# Reset all data
await memory_manager.reset()
if __name__ == "__main__":
asyncio.run(main())
Project Structure
src/
├── mfcs_memory/
│ ├── core/
│ │ ├── base.py # Base manager (connections)
│ │ ├── memory_manager.py # Memory manager (main logic)
│ │ ├── session_manager.py # Session manager (session, chunk, task)
│ │ ├── vector_store.py # Vector store (Qdrant)
│ │ ├── conversation_analyzer.py # Conversation analyzer (LLM)
│ │ └── __init__.py
│ ├── utils/
│ │ ├── config.py # Configuration management
│ │ └── __init__.py
│ └── __init__.py
├── example/ # Example code
├── model/ # Model directory
├── setup.py # Installation config
├── .env.example # Environment file example
└── README.md # Project documentation
Configuration Guide
Required Configuration
MONGO_USER: MongoDB usernameMONGO_PASSWD: MongoDB passwordMONGO_HOST: MongoDB host addressQDRANT_HOST: Qdrant host addressEMBEDDING_MODEL_PATH: Model path for generating text vectorsEMBEDDING_DIM: Vector dimensionOPENAI_API_KEY: OpenAI API keyOPENAI_API_BASE: OpenAI API base URL (Optional)LLM_MODEL: LLM model name
Optional Configuration
MONGO_REPLSET: MongoDB replica set name (if using replica set)QDRANT_PORT: Qdrant port number (default: 6333)MAX_RECENT_HISTORY: Number of recent conversations kept in main table (default: 20)CHUNK_SIZE: Number of conversations stored in each chunk (default: 100)MAX_CONCURRENT_ANALYSIS: Maximum number of concurrent analysis tasks (default: 3)
Contributing
Issues and Pull Requests are welcome!
License
MIT License
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file mfcs_memory-0.1.5.tar.gz.
File metadata
- Download URL: mfcs_memory-0.1.5.tar.gz
- Upload date:
- Size: 16.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.11.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
d5c6c6bec6a705e7365e535a6f76de7e4a460a33d5a1a49a73684559830c9c57
|
|
| MD5 |
e89cf5ab18ee88b3d1255a7ccb10c4e3
|
|
| BLAKE2b-256 |
c83d81f821bb4a982cccd27c6da6f4a3217ad5d9491022fac705613bb7f51fa7
|
File details
Details for the file mfcs_memory-0.1.5-py3-none-any.whl.
File metadata
- Download URL: mfcs_memory-0.1.5-py3-none-any.whl
- Upload date:
- Size: 17.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.11.9
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
5af21f4a3af2233b0a5925ec83834df4150b43546ec9b630272ef333cb076ae4
|
|
| MD5 |
c89b4789e43caec76c980c3cf4479852
|
|
| BLAKE2b-256 |
cd097de2b19e294af5295d4b8de1d568aeebed9564c5d56827d94233d3147ea5
|