Skip to main content

TextEmbed provides a robust and scalable REST API for generating vector embeddings from text. Built for performance and flexibility, it supports various sentence-transformer models, allowing users to easily integrate state-of-the-art NLP techniques into their applications. Whether you need embeddings for search, recommendation, or other NLP tasks, TextEmbed delivers with high efficiency.

Project description

Contributors Issues Apache License 2.0 Downloads Docker Pulls PyPI - Version Reliability Rating Quality Gate Status

TextEmbed - Embedding Inference Server

TextEmbed is a high-throughput, low-latency REST API designed for serving vector embeddings. It supports a wide range of sentence-transformer models and frameworks, making it suitable for various applications in natural language processing.

Features

  • High Throughput & Low Latency: Designed to handle a large number of requests efficiently.
  • Flexible Model Support: Works with various sentence-transformer models.
  • Scalable: Easily integrates into larger systems and scales with demand.
  • Batch Processing: Supports batch processing for better and faster inference.
  • OpenAI Compatible REST API Endpoint: Provides an OpenAI compatible REST API endpoint.
  • Single Line Command Deployment: Deploy multiple models via a single command for efficient deployment.
  • Support for Embedding Formats: Supports binary, float16, and float32 embeddings formats for faster retrieval.

Getting Started

Prerequisites

Ensure you have Python 3.10 or higher installed. You will also need to install the required dependencies.

Installation

  1. Install the required dependencies:

    pip install -U textembed
    
  2. Start the TextEmbed server with your desired models:

    python3 -m textembed.server --models <Model1>, <Model2> --port <Port>
    

    Replace <Model1> and <Model2> with the names of the models you want to use, separated by commas. Replace <Port> with the port number on which you want to run the server.

For more information about the Docker deployment and configuration, please refer to the documentation setup.md.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

textembed-0.0.8.tar.gz (22.7 kB view details)

Uploaded Source

Built Distribution

textembed-0.0.8-py3-none-any.whl (26.5 kB view details)

Uploaded Python 3

File details

Details for the file textembed-0.0.8.tar.gz.

File metadata

  • Download URL: textembed-0.0.8.tar.gz
  • Upload date:
  • Size: 22.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.9.19

File hashes

Hashes for textembed-0.0.8.tar.gz
Algorithm Hash digest
SHA256 b8be1d2ce72c87efe805c83db0c6868bd7416e7bb3a724fe67123f281c9a0aed
MD5 c362c242d3ae5cf026d908c21c8e9094
BLAKE2b-256 3fa42d6d090aeda3003bbbda396a3be8ae1c5db68db38e44b0070f4f608c07f1

See more details on using hashes here.

File details

Details for the file textembed-0.0.8-py3-none-any.whl.

File metadata

  • Download URL: textembed-0.0.8-py3-none-any.whl
  • Upload date:
  • Size: 26.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.9.19

File hashes

Hashes for textembed-0.0.8-py3-none-any.whl
Algorithm Hash digest
SHA256 8034c1f5bf7d705564b14282998aead308e928969930ab4e370c83362b213d0a
MD5 4a7e78ff05dd3ed9bd217459ff4ab1cf
BLAKE2b-256 3ef9ebf165504d63659baa9b9d8eca54139865ba9ab70535a49360b97368b27b

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page