Skip to main content

Chat with your documents locally.

Project description

localrag

localrag is a Python package enabling users to "chat" with their documents using a local Retrieval Augmented Generation (RAG) approach, without needing an external Large Language Model (LLM) provider.

It allows for quick, local, and easy interactions with text data, extracting and generating responses based on the content.

Features

  • Local Processing: Runs entirely on your local machine - no need to send data externally.
  • Customizable: Easy to set up with default models or specify your own.
  • Versatile: Use it for a variety of applications, from automated Q&A systems to data mining.

Prerequisites

Before you install and start using localrag, make sure you meet the following requirements:

Ollama for Local Inference

localrag uses Ollama for local inference, particularly beneficial for macOS users. Ollama allows for easy model serving and inference. To set up Ollama:

Installation

To install localrag, simply use pip:

pip install localrag

Quick Start

Here's a quick example of how you can use localrag to chat with your documents:

Here is an example in test.txt in the docs folder:

I have a dog
import localrag
# can set device to mps or cuda:0 e.g device="mps"
# can set index location e.g index_location="my_index_loc"
my_local_rag = localrag.init()
# Add docs
my_local_rag.add_to_index("./docs")
# Chat with docs
response = my_local_rag.chat("What type of pet do I have?")
print(response.answer)
print(response.source_documents)
# Based on the context you provided, I can determine that you have a dog. Therefore, the type of pet you have is "dog."
# [Document(page_content='I have a dog', metadata={'source': 'docs/test.txt'})]

License

This library is licensed under the Apache 2.0 License. See the LICENSE file.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

localrag-0.1.32.tar.gz (9.1 kB view details)

Uploaded Source

Built Distribution

localrag-0.1.32-py3-none-any.whl (9.6 kB view details)

Uploaded Python 3

File details

Details for the file localrag-0.1.32.tar.gz.

File metadata

  • Download URL: localrag-0.1.32.tar.gz
  • Upload date:
  • Size: 9.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.5

File hashes

Hashes for localrag-0.1.32.tar.gz
Algorithm Hash digest
SHA256 331761d75fb14a119d7d78444dadf78940182ef18cf4327c761b36230ea6b699
MD5 5af093c681f62eeee6e3e6160ad9cb3c
BLAKE2b-256 74fc3e1537052c76161192c312109b0885a256c3c262cdc91046cf50c2f01f3b

See more details on using hashes here.

File details

Details for the file localrag-0.1.32-py3-none-any.whl.

File metadata

  • Download URL: localrag-0.1.32-py3-none-any.whl
  • Upload date:
  • Size: 9.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.5

File hashes

Hashes for localrag-0.1.32-py3-none-any.whl
Algorithm Hash digest
SHA256 11f95e8f930c80314253d24a861491361f8eca347e663f9be54588adc009ca64
MD5 0abee14455447d9769d880390705cd8c
BLAKE2b-256 625589740dde5924e22cfe6a51c55bd0915f7a726def2bf5e8169aeb69bcba50

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page