Skip to main content

Integration of TextEmbed with llama-index for embeddings.

Project description

TextEmbed - Embedding Inference Server

Maintained by Keval Dekivadiya, TextEmbed is licensed under the Apache-2.0 License.

TextEmbed is a high-throughput, low-latency REST API designed for serving vector embeddings. It supports a wide range of sentence-transformer models and frameworks, making it suitable for various applications in natural language processing.

Features

  • High Throughput & Low Latency: Designed to handle a large number of requests efficiently.
  • Flexible Model Support: Works with various sentence-transformer models.
  • Scalable: Easily integrates into larger systems and scales with demand.
  • Batch Processing: Supports batch processing for better and faster inference.
  • OpenAI Compatible REST API Endpoint: Provides an OpenAI compatible REST API endpoint.
  • Single Line Command Deployment: Deploy multiple models via a single command for efficient deployment.
  • Support for Embedding Formats: Supports binary, float16, and float32 embeddings formats for faster retrieval.

Getting Started

Prerequisites

Ensure you have Python 3.10 or higher installed. You will also need to install the required dependencies.

Installation via PyPI

Install the required dependencies:

pip install -U textembed

Start the TextEmbed Server

Start the TextEmbed server with your desired models:

python -m textembed.server --models sentence-transformers/all-MiniLM-L12-v2 --workers 4 --api-key TextEmbed

Example Usage with llama-index

Here's a simple example to get you started with llama-index:

from llama_index.embeddings.textembed import TextEmbedEmbedding

# Initialize the TextEmbedEmbedding class
embed = TextEmbedEmbedding(
    model_name="sentence-transformers/all-MiniLM-L12-v2",
    base_url="http://0.0.0.0:8000/v1",
    auth_token="TextEmbed",
)

# Get embeddings for a batch of texts
embeddings = embed.get_text_embedding_batch(
    [
        "It is raining cats and dogs here!",
        "India has a diverse cultural heritage.",
    ]
)

print(embeddings)

For more information, please read the documentation.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

Built Distribution

File details

Details for the file llama_index_embeddings_textembed-0.1.1.tar.gz.

File metadata

File hashes

Hashes for llama_index_embeddings_textembed-0.1.1.tar.gz
Algorithm Hash digest
SHA256 ee4a78170be5788e7d527a15a8c967c4bf344b731526a3aa466bc939059a1941
MD5 ee0ab68f1da05fde28ebcaad03fd8116
BLAKE2b-256 b23d1b8283c42c48a3cc66e9249a00a8f9035f1fb9b57fadaf53b7182f70f5e5

See more details on using hashes here.

File details

Details for the file llama_index_embeddings_textembed-0.1.1-py3-none-any.whl.

File metadata

File hashes

Hashes for llama_index_embeddings_textembed-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 46474731952e21e0fe1bf09ec1d9cfe52ed0d890f203bf59911420805a6b8cf5
MD5 c4987e9c0505fd7a09150aca9069df86
BLAKE2b-256 5220dbad9b3c0810a59521c7f18d789d45b713e1c8f1ba73d85eae3ab196ff15

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page