Skip to main content

A RAG (Retrieval-Augmented Generation) toolkit with Milvus integration

Project description

RagXO

Export, version and reuse your E2E RAG pipeline everywhere 🚀

PyPI version License: MIT Python 3.8+

RagXO extends the capabilities of traditional RAG (Retrieval-Augmented Generation) systems by providing a unified way to package, version, and deploy your entire RAG pipeline with LLM integration. Export your complete system—including embedding functions, preprocessing steps, vector store, and LLM configurations—into a single, portable artifact.

Features ✨

  • Complete RAG Pipeline: Package your entire RAG system into a versioned artifact
  • LLM Integration: Built-in support for OpenAI models
  • Flexible Embedding: Compatible with any embedding function (Sentence Transformers, OpenAI, etc.)
  • Custom Preprocessing: Chain multiple preprocessing steps
  • Vector Store Integration: Built-in Milvus support
  • System Prompts: Include and version your system prompts

Installation 🛠️

pip install ragxo

Quickstart 🚀

Build a RAG pipeline

export OPENAI_API_KEY=<openai_key> 
from ragxo import Ragxo, Document



ragxo_client = Ragxo(dimension=1536)

def preprocess_text_remove_special_chars(text: str) -> str:
    return re.sub(r'[^a-zA-Z0-9\s]', '', text)

def preprocess_text_lower(text: str) -> str:
    return text.lower()

def get_embeddings(text: str) -> list[float]:
    return openai.embeddings.create(input=text, model="text-embedding-ada-002").data[0].embedding

ragxo_client.add_preprocess(preprocess_text_lower)
ragxo_client.add_preprocess(preprocess_text_remove_special_chars)
ragxo_client.add_embedding_fn(get_embeddings)

ragxo_client.add_system_prompt("You are a helpful assistant that can answer questions about the data provided.")
ragxo_client.add_model(
    "gpt-4o-mini",
    limit=10,
    temperature=0.5,
    max_tokens=1000,
    top_p=1.0,
    frequency_penalty=0.0,
    presence_penalty=0.0
)

ragxo_client.index([
    Document(text="Capital of France is Paris", metadata={"source": "example"}, id=1),
    Document(text="Capital of Germany is Berlin", metadata={"source": "example"}, id=2),
    Document(text="Capital of Italy is Rome", metadata={"source": "example"}, id=3),
])

ragxo_client.export("my_rag_v1.0.0")

# or export to s3
ragxo_client.export("my_rag_v1.0.0", s3_bucket="my_bucket")

Load a RAG pipeline

loaded_ragxo_client = Ragxo.load("my_rag_v1.0.0")

vector_search_results = loaded_ragxo_client.query("What is the capital of France?")

llm_response = loaded_ragxo_client.generate_llm_response(
    "What is the capital of France?")

print(llm_response.choices[0].message.content)

Usage Guide 📚

Import

from ragxo import Ragxo, Document

ragxo_client = Ragxo(dimension=768)

Adding Preprocessing Steps

import re

def remove_special_chars(text: str) -> str:
    return re.sub(r'[^a-zA-Z0-9\s]', '', text)

def lowercase(text: str) -> str:
    return text.lower()

ragxo_client.add_preprocess(remove_special_chars)
ragxo_client.add_preprocess(lowercase)

Custom Embedding Functions

# Using SentenceTransformers
from sentence_transformers import SentenceTransformer
model = SentenceTransformer('all-MiniLM-L6-v2')

def get_embeddings(text: str) -> list[float]:
    return model.encode(text).tolist()

ragxo.add_embedding_fn(get_embeddings)

# Or using OpenAI
from openai import OpenAI
client = OpenAI()

def get_openai_embeddings(text: str) -> list[float]:
    response = client.embeddings.create(
        input=text,
        model="text-embedding-ada-002"
    )
    return response.data[0].embedding

ragxo.add_embedding_fn(get_openai_embeddings)

Creating Documents

from ragxo import Document

doc = Document(
    text="Your document content here",
    metadata={"source": "wiki", "category": "science"},
    id=1
)

ragxo_client.index([doc])

LLM Configuration

# Set system prompt
ragxo_client.add_system_prompt("""
You are a helpful assistant. Use the provided context to answer questions accurately.
If you're unsure about something, please say so.
""")

# Set LLM model
ragxo_client.add_model("gpt-4")

Export and Load

# Export your RAG pipeline
ragxo_client.export("rag_pipeline_v1")

# Load it elsewhere
loaded_ragxo_client = Ragxo.load("rag_pipeline_v1")

Best Practices 💡

  1. Version Your Exports: Use semantic versioning for your exports:
ragxo.export("my_rag_v1.0.0")
  1. S3: Use S3 to store your exports
export AWS_ACCESS_KEY_ID=your_access_key
export AWS_SECRET_ACCESS_KEY=your_secret_key
ragxo_client.export("my_rag_v1.0.0", s3_bucket="my_bucket")

License 📝

This project is licensed under the MIT License - see the LICENSE file for details.

Contributing 🤝

Contributions are welcome! Please feel free to submit a Pull Request.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ragxo-0.1.11.tar.gz (6.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

ragxo-0.1.11-py3-none-any.whl (6.8 kB view details)

Uploaded Python 3

File details

Details for the file ragxo-0.1.11.tar.gz.

File metadata

  • Download URL: ragxo-0.1.11.tar.gz
  • Upload date:
  • Size: 6.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.5 CPython/3.13.1 Darwin/24.1.0

File hashes

Hashes for ragxo-0.1.11.tar.gz
Algorithm Hash digest
SHA256 1025bac401848c1b0b527c2ad691f98c140389a96b5b52831167bdc89a9b460b
MD5 9cff9b8cc30c58b5647dd4e84aeb796d
BLAKE2b-256 d39e548e2fad82cee68dcff25ed4054e7163179f2baa0cc532916d332e9a8f31

See more details on using hashes here.

File details

Details for the file ragxo-0.1.11-py3-none-any.whl.

File metadata

  • Download URL: ragxo-0.1.11-py3-none-any.whl
  • Upload date:
  • Size: 6.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.5 CPython/3.13.1 Darwin/24.1.0

File hashes

Hashes for ragxo-0.1.11-py3-none-any.whl
Algorithm Hash digest
SHA256 c2b756274caa8468a5b8187f75589964c21dc7d1c7d9c933fdd39f774be0cd80
MD5 74f79bda7cb83d4cb5a396f98b4fc9fe
BLAKE2b-256 e894ea4f65339ab1ac85065a90fbfc8ddf728e86e32c6ba0135f60969bd60ccb

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page