Skip to main content

AutoRAG is a flexible and scalable solution for building Retrieval-Augmented Generation (RAG) systems.

Project description

AutoRAG

Powering seamless retrieval and generation workflows for our internal AI systems

Python Version PyPI Version Code Formatter Code Linter Code Checker Code Coverage

Overview

AutoRAG is a flexible and scalable solution for building Retrieval-Augmented Generation (RAG) systems.

This SDK provides out-of-the-box functionality for creating and managing retrieval-augmented generation workflows, offering a modular, highly-configurable interface. It supports multiple vector stores and leverages http clients like httpx for handling requests, ensuring seamless integration.

Features

  • Modular architecture: The SDK allows you to swap, extend, or customize components like retrieval models, vector stores, and response generation strategies.
  • High scalability: Built to handle large-scale data retrieval and generation, enabling robust, production-ready applications.
  • Celery for dependency injection: Efficient background tasks with support for distributed task execution.
  • Multi-flow support: Easily integrate various vector databases (ex: Qdrant, Azure AI Search) with various language models providers (ex: OpenAI, vLLM, Ollama) using standardized public methods for seamless development.

Installation

  1. Create a virtual environment, we recommend Miniconda for environment management:
    conda create -n autorag python=3.12
    conda activate autorag
    
  2. Install the package:
    pip install autonomize-autorag
    

To install with optional dependencies like Qdrant, Huggingface, OpenAI, Modelhub, etc., refer to the Installation Guide.

Usage

The full set of examples can be found in examples directory.

Sync Usage

import os
from autorag.language_models import OpenAILanguageModel

llm = OpenAILanguageModel(
    api_key=os.environ.get("OPENAI_API_KEY"),
)

generation = llm.generate(
    message=[{"role": "user", "content": "What is attention in ML?"}],
    model="gpt-4o"
)

Async Usage

Simply use sync methods with a prefix and use await for each call. Example: client.generate(...) becomes await client.agenerate(...) and everything else remains the same.

import os
from autorag.language_models import OpenAILanguageModel

llm = OpenAILanguageModel(
    api_key=os.environ.get("OPENAI_API_KEY"),
)

generation = await llm.agenerate(
    message=[{"role": "user", "content": "What is attention in ML?"}],
    model="gpt-4o"
)

Contribution

To contribute in our AutoRAG SDK, please refer to our Contribution Guidelines.

License

Copyright (C) Autonomize AI - All Rights Reserved

The contents of this repository cannot be copied and/or distributed without the explicit permission from Autonomize.ai

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

autonomize_autorag-0.1.83.tar.gz (53.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

autonomize_autorag-0.1.83-py3-none-any.whl (84.3 kB view details)

Uploaded Python 3

File details

Details for the file autonomize_autorag-0.1.83.tar.gz.

File metadata

  • Download URL: autonomize_autorag-0.1.83.tar.gz
  • Upload date:
  • Size: 53.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.2.1 CPython/3.12.12 Linux/6.11.0-1018-azure

File hashes

Hashes for autonomize_autorag-0.1.83.tar.gz
Algorithm Hash digest
SHA256 46b9d054ef5c70ed4544e0993549477e8dd422735d3e71e3a777da6838d70850
MD5 f66df58b2afdd11d61dcd0dc2190d2e1
BLAKE2b-256 1746440d8f439cd02b47d19d0c7c66def1a52eee1c11b6717137f4b1df9f6c98

See more details on using hashes here.

File details

Details for the file autonomize_autorag-0.1.83-py3-none-any.whl.

File metadata

  • Download URL: autonomize_autorag-0.1.83-py3-none-any.whl
  • Upload date:
  • Size: 84.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.2.1 CPython/3.12.12 Linux/6.11.0-1018-azure

File hashes

Hashes for autonomize_autorag-0.1.83-py3-none-any.whl
Algorithm Hash digest
SHA256 ec06a735c21363f08866a054431f5ab2106022d1035ef454a35af9879fe8847e
MD5 c109838d9f59a3c02afc49c11bc14211
BLAKE2b-256 e41eb5f679cf0270d4831aea836caebef35116b05083f28ce05f47153eafac98

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page