Skip to main content

LLM framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data.

Project description

Green logo of a stylized white 'H' with the text 'Deepstack, by khulnasoft. Deepstack 2.0 is live 🎉' Abstract green and yellow diagrams in the background.
CI/CD Tests code style - Black types - Mypy Coverage Status
Docs Website
Package PyPI PyPI - Downloads PyPI - Python Version Conda Version GitHub License Compliance
Meta Discord Twitter Follow

Deepstack is an end-to-end LLM framework that allows you to build applications powered by LLMs, Transformer models, vector search and more. Whether you want to perform retrieval-augmented generation (RAG), document search, question answering or answer generation, Deepstack can orchestrate state-of-the-art embedding models and LLMs into pipelines to build end-to-end NLP applications and solve your use case.

Installation

The simplest way to get Deepstack is via pip:

pip install deepstack-ai

Deepstack supports multiple installation methods including Docker images. For a comprehensive guide please refer to the documentation.

Documentation

If you're new to the project, check out "What is Deepstack?" then go through the "Get Started Guide" and build your first LLM application in a matter of minutes. Keep learning with the tutorials. For more advanced use cases, or just to get some inspiration, you can browse our Deepstack recipes in the Cookbook.

At any given point, hit the documentation to learn more about Deepstack, what can it do for you and the technology behind.

Features

[!IMPORTANT] You are currently looking at the readme of Deepstack 2.0. We are still maintaining Deepstack 1.x to give everyone enough time to migrate to 2.0. Switch to Deepstack 1.x here.

  • Technology agnostic: Allow users the flexibility to decide what vendor or technology they want and make it easy to switch out any component for another. Deepstack allows you to use and compare models available from OpenAI, Cohere and Hugging Face, as well as your own local models or models hosted on Azure, Bedrock and SageMaker.
  • Explicit: Make it transparent how different moving parts can “talk” to each other so it's easier to fit your tech stack and use case.
  • Flexible: Deepstack provides all tooling in one place: database access, file conversion, cleaning, splitting, training, eval, inference, and more. And whenever custom behavior is desirable, it's easy to create custom components.
  • Extensible: Provide a uniform and easy way for the community and third parties to build their own components and foster an open ecosystem around Deepstack.

Some examples of what you can do with Deepstack:

  • Build retrieval augmented generation (RAG) by making use of one of the available vector databases and customizing your LLM interaction, the sky is the limit 🚀
  • Perform Question Answering in natural language to find granular answers in your documents.
  • Perform semantic search and retrieve documents according to meaning.
  • Build applications that can make complex decisions making to answer complex queries: such as systems that can resolve complex customer queries, do knowledge search on many disconnected resources and so on.
  • Scale to millions of docs using retrievers and production-scale components.
  • Use off-the-shelf models or fine-tune them to your data.
  • Use user feedback to evaluate, benchmark, and continuously improve your models.

[!TIP]

Are you looking for a managed solution that benefits from Deepstack? khulnasoft Cloud is our fully managed, end-to-end platform to integrate LLMs with your data, which uses Deepstack for the LLM pipelines architecture.

Telemetry

Deepstack collects anonymous usage statistics of pipeline components. We receive an event every time these components are initialized. This way, we know which components are most relevant to our community.

Read more about telemetry in Deepstack or how you can opt out in Deepstack docs.

🖖 Community

If you have a feature request or a bug report, feel free to open an issue in Github. We regularly check these and you can expect a quick response. If you'd like to discuss a topic, or get more general advice on how to make Deepstack work for your project, you can start a thread in Github Discussions or our Discord channel. We also check 𝕏 (Twitter) and Stack Overflow.

Contributing to Deepstack

We are very open to the community's contributions - be it a quick fix of a typo, or a completely new feature! You don't need to be a Deepstack expert to provide meaningful improvements. To learn how to get started, check out our Contributor Guidelines first.

There are several ways you can contribute to Deepstack:

[!TIP] 👉 Check out the full list of issues that are open to contributions

Who Uses Deepstack

Here's a list of projects and companies using Deepstack. Want to add yours? Open a PR, add it to the list and let the world know that you use Deepstack!

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

deepstack_ai-2.2.0rc0.tar.gz (206.7 kB view details)

Uploaded Source

Built Distribution

deepstack_ai-2.2.0rc0-py3-none-any.whl (320.9 kB view details)

Uploaded Python 3

File details

Details for the file deepstack_ai-2.2.0rc0.tar.gz.

File metadata

  • Download URL: deepstack_ai-2.2.0rc0.tar.gz
  • Upload date:
  • Size: 206.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.12.3

File hashes

Hashes for deepstack_ai-2.2.0rc0.tar.gz
Algorithm Hash digest
SHA256 f07723f10a3eaae2a54e30bf367fbc1f6ed9487c29e0d4515b1cee21ce2974ae
MD5 ef4cb142f6621ffafbf24e24756221e5
BLAKE2b-256 7865eaa5e6e001fefe4fadf6c74d1a29c99218291a20ff0b25ddffafe33b8c13

See more details on using hashes here.

File details

Details for the file deepstack_ai-2.2.0rc0-py3-none-any.whl.

File metadata

File hashes

Hashes for deepstack_ai-2.2.0rc0-py3-none-any.whl
Algorithm Hash digest
SHA256 1ea997a7787ac3ba6583b2ec6ad9a32c7b14fe8596dcc5e610432ba9c49dfd04
MD5 27f9018ae0af040a651a476e42e23651
BLAKE2b-256 a82b0a41973c26f9ccfca1b3897971422042013ce61dadcfd33c8c86de45171b

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page