Skip to main content

LLM Powered Advanced RAG Application

Project description

Open In Colab

pyLLMSearch - Advanced RAG

Documentation

The purpose of this package is to offer an advanced question-answering (RAG) system with a simple YAML-based configuration that enables interaction with a collection of local documents. Special attention is given to improvements in various components of the system in addition to basic LLM-based RAGs - better document parsing, hybrid search, HyDE, chat history, deep linking, re-ranking, the ability to customize embeddings, and more. The package is designed to work with custom Large Language Models (LLMs) – whether from OpenAI or installed locally.

Interaction with the package is supported through the built-in frontend, or by exposing an MCP server, allowing clients like Cursor, Windsurf or VSCode GH Copilot to interact with the RAG system.

Features

  • Fast parsing and embedding of medium size document bases (tested on up to few gigabytes of markdown and pdfs)

  • Incremental updates for new documents, without a need to re-index the entire document base.

  • Supported document formats

    • Build-in parsers:
      • .md - Divides files based on logical components such as headings, subheadings, and code blocks. Supports additional features like cleaning image links, adding custom metadata, and more.
      • .pdf - MuPDF-based parser.
      • .docx - custom parser, supports nested tables.
    • Other common formats are supported by Unstructured pre-processor:
      • List of formats see here.
  • FastAPI based API + MCP server, allowing communicating with RAG via any mcp client, including VSCode/Windsurf/Cursor and others.

  • Deep linking into document sections - jump to an individual PDF page or a header in a markdown file.

  • Allows interaction with embedded documents, internally supporting the following models and methods (including locally hosted):

    • OpenAI compatible models and APIs.
    • HuggingFace models.
  • Interoperability with LiteLLM + Ollama via OpenAI API, supporting hundreds of different models (see Model configuration for LiteLLM)

  • SSE MCP Server enabling interface with popular MCP clients.

  • Hybrid search and Reranking

    • Dense embeddings from a folder of documents and stores them in a vector database (ChromaDB).

    • The following embedding models are supported:

      • Hugging Face embeddings.
      • Sentence-transformers-based models.
      • Instructor-based models.
      • OpenAI embeddings.
    • Sparse embeddings using SPLADE (https://github.com/naver/splade) to enable hybrid search (sparse + dense).

    • Supports the "Retrieve and Re-rank" strategy for semantic search, see here.

      • Besides the originally ms-marco-MiniLM cross-encoder, more modern bge-reranker-v2-m3 and zerank-2 is supported.
  • Support for table parsing via open-source gmft (https://github.com/conjuncts/gmft) or Azure Document Intelligence.

  • Optional support for image parsing using Gemini API.

  • Supports HyDE (Hypothetical Document Embeddings) - see here.

    • WARNING: Enabling HyDE (via config OR webapp) can significantly alter the quality of the results. Please make sure to read the paper before enabling.
    • From my own experiments, enabling HyDE significantly boosts quality of the output on a topics where user can't formulate the quesiton using domain specific language of the topic - e.g. when learning new topics.
  • Support for multi-querying, inspired by RAG Fusion - https://towardsdatascience.com/forget-rag-the-future-is-rag-fusion-1147298d8ad1

    • When multi-querying is turned on (either config or webapp), the original query will be replaced by 3 variants of the same query, allowing to bridge the gap in the terminology and "offer different angles or perspectives" according to the article.
  • Supprts optional chat history with question contextualization

  • Other features

    • Simple web interfaces.
    • Ability to save responses to an offline database for future analysis.

Demo

Demo

Documentation

Browse Documentation

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pyllmsearch-0.9.5.tar.gz (4.0 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

pyllmsearch-0.9.5-py3-none-any.whl (76.5 kB view details)

Uploaded Python 3

File details

Details for the file pyllmsearch-0.9.5.tar.gz.

File metadata

  • Download URL: pyllmsearch-0.9.5.tar.gz
  • Upload date:
  • Size: 4.0 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.25

File hashes

Hashes for pyllmsearch-0.9.5.tar.gz
Algorithm Hash digest
SHA256 de5a5e68dfac7fff3feb6ef399b31cbe232b6506e9e0374b6487496b881c1d15
MD5 12b49d5421cb134195e0463f8fa0657b
BLAKE2b-256 ea33b83ba1810a3da76cef2da5f93cebcb7c8f7be5622351ca46c8baf8171c64

See more details on using hashes here.

File details

Details for the file pyllmsearch-0.9.5-py3-none-any.whl.

File metadata

  • Download URL: pyllmsearch-0.9.5-py3-none-any.whl
  • Upload date:
  • Size: 76.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.25

File hashes

Hashes for pyllmsearch-0.9.5-py3-none-any.whl
Algorithm Hash digest
SHA256 949bf9f76a496ed00815e48ea6308bc09785e02d1ae3664ccc09bd023fbe7da1
MD5 995ecf013e6ad4a4d42f51457c1f27be
BLAKE2b-256 33d5ee46ded8c3a7fc2c77614fa7d064277afc55c0ce91091e09079a6738fdd1

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page