A RAG system for creating knowledge bases from different document formats
Project description
Ragora
Build smarter, grounded, and transparent AI with Ragora.
Ragora is an open-source framework for building Retrieval-Augmented Generation (RAG) systems that connect your language models to real, reliable knowledge. It provides a clean, composable interface for managing knowledge bases, document retrieval, and grounding pipelines, so your AI can reason with context instead of guesswork.
The name Ragora blends RAG with the ancient Greek Agora, the public square where ideas were exchanged, debated, and refined. In the same spirit, Ragora is the meeting place of data and dialogue, where your information and your AI come together to think.
✨ Key Features
- 📄 Specialized Document Processing: Native support for LaTeX parsing and email handling with more formats coming
- 🏗️ Clean Architecture: Three-layer design (DatabaseManager → VectorStore → Retriever) for maintainability
- 🔍 Flexible Search: Vector, keyword, and hybrid search modes for optimal retrieval
- 🧩 Composable Components: Use high-level APIs or build custom pipelines with low-level components
- ⚡ Performance Optimized: Batch processing, GPU acceleration, and efficient vector search with Weaviate
- 🔒 Privacy-First: Run completely local with sentence-transformers and Weaviate
🚀 Installation
pip install ragora
Prerequisites
You need a Weaviate instance running. Download the pre-configured Ragora database server:
# Download from GitHub releases
wget https://github.com/Vahidlari/aiApps/releases/download/v<x.y.z>/database_server-<x.y.z>.tar.gz
# Extract and start
tar -xzf database_server-<x.y.z>.tar.gz
cd database-server
./database-manager.sh start
Update <x.y.z> with the actual package version- For example use 1.0.0 for version v1.0.0.
The database server is a zero-dependency solution (only requires Docker) that works on Windows, macOS, and Linux.
Document Processing
Process LaTeX documents with specialized handling:
from ragora.core import DocumentPreprocessor, DataChunker
# Parse LaTeX with citations
preprocessor = DocumentPreprocessor()
document = preprocessor.parse_latex(
"paper.tex",
bibliography_path="references.bib"
)
# Chunk with configurable size and overlap using new API
from ragora import DataChunker, ChunkingContextBuilder
chunker = DataChunker()
context = ChunkingContextBuilder().for_document().build()
chunks = chunker.chunk(document.content, context)
🔍 Search Modes
Ragora supports three search strategies:
from ragora import SearchStrategy
# Semantic search (best for conceptual queries)
results = kbm.search("explain machine learning", strategy=SearchStrategy.SIMILAR)
# Keyword search (best for exact terms)
results = kbm.search("Schrödinger equation", strategy=SearchStrategy.KEYWORD)
# Hybrid search (recommended - combines both)
results = kbm.search("neural networks", strategy=SearchStrategy.HYBRID, alpha=0.7)
🎯 Use Cases
- 📖 Academic Research: Build knowledge bases from scientific papers and LaTeX documents
- 📝 Documentation Search: Create searchable knowledge bases from technical documentation
- 🤖 AI Assistants: Ground LLM responses in your specific domain knowledge
- 💬 Question Answering: Build Q&A systems over your document collections
- 🔬 Literature Review: Efficiently search and synthesize information from research papers
📖 Documentation & Examples
- Getting Started Guide: Detailed installation and setup guide
- API Reference: Complete API documentation
- Examples Directory: Working code examples
basic_usage.py: Basic usage examples and getting startedadvanced_usage.py: Advanced features and custom pipelinesemail_usage_examples.py: Email integration examples
📊 Requirements
- Python: 3.11 or higher
- Weaviate: 1.22.0 or higher (for vector storage)
- Dependencies: See requirements.txt
🤝 Contributing
We welcome contributions! Please see our Contributing Guidelines for:
- Setting up your development environment
- Code style and standards
- Writing tests
- Submitting pull requests
📄 License
This project is licensed under the MIT License - see the LICENSE file for details.
🔗 Links
- Repository: github.com/vahidlari/aiapps
- Issues: GitHub Issues
- Discussions: GitHub Discussions
📮 Contact
For questions, feedback, or collaboration opportunities:
- Open an issue on GitHub
- Start a discussion in GitHub Discussions
- Contact the maintainers directly
Build smarter, grounded, and transparent AI with Ragora.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file ragora-1.2.0.tar.gz.
File metadata
- Download URL: ragora-1.2.0.tar.gz
- Upload date:
- Size: 165.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
47cf74139b05d214f1b7d2b346a684c43ecaa03e65078b740e8cdf1afd8b727e
|
|
| MD5 |
d6653ca5d912d70198ef782be75ba8c2
|
|
| BLAKE2b-256 |
21c1381971c3f05ee6be70737f5a2646e3dfce74d0b4c19e978704b337809179
|
File details
Details for the file ragora-1.2.0-py3-none-any.whl.
File metadata
- Download URL: ragora-1.2.0-py3-none-any.whl
- Upload date:
- Size: 89.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
f065f47080c29366077c2967becf8e61c317ec81e37919a0873ce5530a536299
|
|
| MD5 |
7978816b9ca6723567df1ef2ec2c6f23
|
|
| BLAKE2b-256 |
f041ac75ea1e2686f430343752801a046860c315dec5251db1d9ca93fffd0ef2
|