MemoryScope is a powerful and flexible long term memory system for LLM chatbots. It consists of a memory database and three customizable system operations, which can be flexibly combined to provide robust long term memory services for your LLM chatbot.
Project description
English | 中文
MemoryScope
Equip your LLM chatbot with a powerful and flexible long term memory system.
📰 News
- [2024-09-10] We release MemoryScope v0.1.1.0 now, which is also available in PyPI!
🌟 What is MemoryScope?
MemoryScope provides LLM chatbots with powerful and flexible long-term memory capabilities, offering a framework for building such abilities. It can be applied to scenarios like personal assistants and emotional companions, continuously learning through long-term memory to remember users' basic information as well as various habits and preferences. This allows users to gradually experience a sense of "understanding" when using the LLM.
Demo
Framework
💾 Memory Database: MemoryScope is equipped with a vector database (default is ElasticSearch) to store all memory fragments recorded in the system.
🔧 Worker Library: MemoryScope atomizes the capabilities of long-term memory into individual workers, including over 20 workers for tasks such as query information filtering, observation extraction, and insight updating.
🛠️ Operation Library: Based on the worker pipeline, it constructs the operations for memory services, realizing key capabilities such as memory retrieval and memory consolidation.
- Memory Retrieval: Upon arrival of a user query, this operation returns the semantically related memory pieces and/or those from the corresponding time if the query involves reference to time.
- Memory Consolidation: This operation takes in a batch of user queries and returns important user information extracted from the queries as consolidated observations to be stored in the memory database.
- Reflection and Re-consolidation: At regular intervals, this operation performs reflection upon newly recorded observations to form and update insights. Then, memory re-consolidation is performed to ensure contradictions and repetitions among memory pieces are properly handled.
⚙️ Best Practices:
- Based on the core capabilities of long-term memory, MemoryScope has implemented a dialogue interface (API) with long-term memory and a command-line dialogue practice (CLI) with long-term memory.
- MemoryScope combines currently popular agent frameworks (AutoGen, AgentScope) to provide best practices.
Main Features
⚡ Low response-time (RT) for the user:
- Backend operations (Memory Consolidation, Reflection and Re-consolidation) are decoupled from the frontend operation (Memory Retrieval) in the system.
- While backend operations are usually (and are recommended to be) queued or executed at regular intervals, the system's response time (RT) for the user depends solely on the frontend operation, which is only ~500ms.
🌲 Hierarchical and coherent memory:
- The memory pieces stored in the system are in a hierarchical structure, with insights being the high level information from the aggregation of similarly-themed observations.
- Contradictions and repetitions among memory pieces are handled periodically to ensure coherence of memory.
- Fictitious contents from the user are filtered out to avoid hallucinations by the LLM.
⏰ Time awareness:
- The system is time sensitive when performing both Memory Retrieval and Memory Consolidation. Therefore, it can retrieve accurate relevant information when the query involves reference to time.
💼 Supported Model API
Backend | Task | Some Supported Models |
---|---|---|
openai_backend | Generation | gpt-4o, gpt-4o-mini, gpt-4, gpt-3.5-turbo |
Embedding | text-embedding-ada-002, text-embedding-3-large, text-embedding-3-small | |
dashscope_backend | Generation | qwen-max, qwen-plus, qwen-plus, qwen2-72b-instruct |
Embedding | text-embedding-v1, text-embedding-v2 | |
Reranker | gte-rerank |
In the future, we will support more model interfaces and local deployment of LLM and embedding services.
🚀 Installation
For installation, please refer to Installation.md.
🍕 Quick Start
💡 Contribute
Contributions are always encouraged!
We highly recommend install pre-commit hooks in this repo before committing pull requests. These hooks are small house-keeping scripts executed every time you make a git commit, which will take care of the formatting and linting automatically.
pip install -e .
pre-commit install
Please refer to our Contribution Guide for more details.
📖 Citation
Reference to cite if you use MemoryScope in a paper:
@software{MemoryScope,
author = {Li Yu and
Tiancheng Qin and
Qingxu Fu and
Sen Huang and
Xianzhe Xu and
Zhaoyang Liu and
Boyin Liu},
month = {09},
title = {{MemoryScope}},
url = {https://github.com/modelscope/MemoryScope},
year = {2024}
}
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file memoryscope-0.1.1.0.tar.gz
.
File metadata
- Download URL: memoryscope-0.1.1.0.tar.gz
- Upload date:
- Size: 114.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.12.4
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 9d5dbf5f28d68ae740c4123be6f512a50c6a9c2fca77ec9095ab0b9e911e075c |
|
MD5 | 755efc43354f50af42b630ddb2538906 |
|
BLAKE2b-256 | a3f4c7a1729fc411090bb735950ab7e9f3d92dbca84c4aff835eeea865c0980f |
File details
Details for the file memoryscope-0.1.1.0-py3-none-any.whl
.
File metadata
- Download URL: memoryscope-0.1.1.0-py3-none-any.whl
- Upload date:
- Size: 154.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.12.4
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | ef81b25f959badb55e7136097e84b5581c9fda9d5ffe498cd15f90dd4f33f14c |
|
MD5 | 61c6a39138812f2d1effb939a681a84b |
|
BLAKE2b-256 | 1a470c8b2a9bf7db9463a65c72c7bb3a472684b9f7fd260051f3ed03a78a0d25 |