An integration of Qdrant ANN vector database backend with txtai
Project description
qdrant-txtai
txtai simplifies building AI-powered semantic search applications using Transformers. It leverages the neural embeddings and their properties to encode high-dimensional data in a lower-dimensional space and allows to find similar objects based on their embeddings' proximity.
Implementing such application in real-world use cases requires storing the embeddings in an efficient way though, namely in a vector database like Qdrant. It offers not only a powerful engine for neural search, but also allows setting up a whole cluster if your data does not fit a single machine anymore. It is production grade and can be launched easily with Docker.
Combining the easiness of txtai with Qdrant's performance enables you to build production-ready semantic search applications way faster than before.
Installation
The library might be installed with pip as following:
pip install qdrant-txtai
Usage
Running the txtai application with Qdrant as a vector storage requires launching a Qdrant instance. That might be done easily with Docker:
docker run -p 6333:6333 -p:6334:6334 qdrant/qdrant:v1.0.1
Running the txtai application might be done either programmatically or by providing configuration in a YAML file.
Programmatically
from txtai.embeddings import Embeddings
embeddings = Embeddings({
"path": "sentence-transformers/all-MiniLM-L6-v2",
"backend": "qdrant_txtai.ann.qdrant.Qdrant",
})
embeddings.index([(0, "Correct", None), (1, "Not what we hoped", None)])
result = embeddings.search("positive", 1)
print(result)
Via YAML configuration
# app.yml
embeddings:
path: sentence-transformers/all-MiniLM-L6-v2
backend: qdrant_txtai.ann.qdrant.Qdrant
CONFIG=app.yml uvicorn "txtai.api:app"
curl -X GET "http://localhost:8000/search?query=positive"
Configuration properties
qdrant-txtai allows you to configure both the connection details, and some internal properties of the vector collection which may impact both speed and accuracy. Please refer to Qdrant docs if you are interested in the meaning of each property.
The example below presents all the available options, if we connect to Qdrant server:
embeddings:
path: sentence-transformers/all-MiniLM-L6-v2
backend: qdrant_txtai.ann.qdrant.Qdrant
metric: l2 # allowed values: l2 / cosine / ip
qdrant:
url: qdrant.host
port: 6333
grpc_port: 6334
prefer_grpc: true
collection: CustomCollectionName
https: true # for Qdrant Cloud
api_key: XYZ # for Qdrant Cloud
hnsw:
m: 8
ef_construct: 256
full_scan_threshold:
ef_search: 512
Local in-memory/disk-persisted mode
Qdrant Python client, from version 1.1.1, supports local in-memory/disk-persisted mode. That's a good choice for any test scenarios and quick experiments in which you do not plan to store lots of vectors. In such a case spinning a Docker container might be even not required.
In-memory storage
In case you want to have a transient storage, for example in case of automated tests launched during your CI/CD pipeline, using Qdrant Local mode with in-memory storage might be a preferred option.
embeddings:
path: sentence-transformers/all-MiniLM-L6-v2
backend: qdrant_txtai.ann.qdrant.Qdrant
metric: l2 # allowed values: l2 / cosine / ip
qdrant:
location: ':memory:'
prefer_grpc: true
On disk storage
However, if you prefer to keep the vectors between different runs of your application, it might be better to use on disk storage and pass the path that should be used to persist the data.
embeddings:
path: sentence-transformers/all-MiniLM-L6-v2
backend: qdrant_txtai.ann.qdrant.Qdrant
metric: l2 # allowed values: l2 / cosine / ip
qdrant:
path: '/home/qdrant/storage_local'
prefer_grpc: true
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for qdrant_txtai-1.0.1-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | de271c1cfc039f22f53f0cd5dccbdacad84ea0073de339abdc603ad96f626020 |
|
MD5 | 91083e59511834d8ddc91952dc9f0b9e |
|
BLAKE2b-256 | a715812bfb7af70ad466e8adc6601cfbd23de41a201ae86a3b538f6e1f6dc2ec |