AutoRAG is a flexible and scalable solution for building Retrieval-Augmented Generation (RAG) systems.
Project description
AutoRAG
Powering seamless retrieval and generation workflows for our internal AI systems
Overview
AutoRAG is a flexible and scalable solution for building Retrieval-Augmented Generation (RAG) systems.
This SDK provides out-of-the-box functionality for creating and managing retrieval-augmented generation workflows, offering a modular, highly-configurable interface. It supports multiple vector stores and leverages http clients like httpx for handling requests, ensuring seamless integration.
Features
- Modular architecture: The SDK allows you to swap, extend, or customize components like retrieval models, vector stores, and response generation strategies.
- High scalability: Built to handle large-scale data retrieval and generation, enabling robust, production-ready applications.
- Celery for dependency injection: Efficient background tasks with support for distributed task execution.
- Multi-flow support: Easily integrate various vector databases (ex: Qdrant, Azure AI Search) with various language models providers (ex: OpenAI, vLLM, Ollama) using standardized public methods for seamless development.
Installation
- Create a virtual environment, we recommend Miniconda for environment management:
conda create -n autorag python=3.12 conda activate autorag
- Install the package:
pip install autonomize-autorag
To install with optional dependencies like Qdrant, Huggingface, OpenAI, Modelhub, etc., refer to the Installation Guide.
Usage
The full set of examples can be found in examples directory.
Sync Usage
import os
from autorag.language_models.openai import OpenAILanguageModel
llm = OpenAILanguageModel(
api_key=os.environ.get("OPENAI_API_KEY"),
)
generation = llm.generate(
message=[{"role": "user", "content": "What is attention in ML?"}],
model="gpt-4o"
)
Async Usage
Simply use sync methods with a
prefix and use await
for each call. Example: client.generate(...)
becomes await client.agenerate(...)
and everything else remains the same.
import os
from autorag.language_models.openai import OpenAILanguageModel
llm = OpenAILanguageModel(
api_key=os.environ.get("OPENAI_API_KEY"),
)
generation = await llm.agenerate(
message=[{"role": "user", "content": "What is attention in ML?"}],
model="gpt-4o"
)
Contribution
To contribute in our AutoRAG SDK, please refer to our Contribution Guidelines.
License
Copyright (C) Autonomize AI - All Rights Reserved
The contents of this repository cannot be copied and/or distributed without the explicit permission from Autonomize.ai
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file autonomize_autorag-0.1.20.tar.gz
.
File metadata
- Download URL: autonomize_autorag-0.1.20.tar.gz
- Upload date:
- Size: 22.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.4 CPython/3.12.7 Linux/6.5.0-1025-azure
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 1c36b30dbc3d6ea5225603ec379d736c061bc54cf12b80241588305c32a658e9 |
|
MD5 | 2fb1748701e914ffae78582b41fc4547 |
|
BLAKE2b-256 | 62890425661257dde09e158c5121a9504ada5524deefecd737e347de89d69bdc |
File details
Details for the file autonomize_autorag-0.1.20-py3-none-any.whl
.
File metadata
- Download URL: autonomize_autorag-0.1.20-py3-none-any.whl
- Upload date:
- Size: 37.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.4 CPython/3.12.7 Linux/6.5.0-1025-azure
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | f990b092e7654aaeb1da0c1f33874057b0edca902bb259fcd0bc0b09516499eb |
|
MD5 | dbd58906e09fc403364153660e35fdef |
|
BLAKE2b-256 | 3f42c4a80a918feee402ce64588d9e7897340eb5223fc5e6cec6f0e56d904219 |