Skip to main content

TextEmbed provides a robust and scalable REST API for generating vector embeddings from text. Built for performance and flexibility, it supports various sentence-transformer models, allowing users to easily integrate state-of-the-art NLP techniques into their applications. Whether you need embeddings for search, recommendation, or other NLP tasks, TextEmbed delivers with high efficiency.

Project description

TextEmbed - Embedding Inference Server

TextEmbed is a high-throughput, low-latency REST API designed for serving vector embeddings. It supports a wide range of sentence-transformer models and frameworks, making it suitable for various applications in natural language processing.

Features

  • High Throughput & Low Latency: Designed to handle a large number of requests efficiently.
  • Flexible Model Support: Works with various sentence-transformer models.
  • Scalable: Easily integrates into larger systems and scales with demand.
  • Batch Processing: Supports batch processing for better and faster inference.
  • OpenAI Compatible REST API Endpoint: Provides an OpenAI compatible REST API endpoint.

Getting Started

Prerequisites

Ensure you have Python 3.11 or higher installed. You will also need to install the required dependencies.

Installation

  1. Install the required dependencies:

    pip install -r requirements.txt
    
  2. Start the TextEmbed server with your desired model:

    python3 -m textembed.server --model <Model Name>
    

    Replace <Model Name> with the name of the model you want to use.

  3. For more information and additional options, run:

    python3 -m textembed.server --help
    

Running with Docker (Recommended)

You can also run TextEmbed using Docker. The Docker image is available on Docker Hub.

docker run kevaldekivadiya/textembed:latest --help

This command will show the help message for the TextEmbed server, detailing the available options and usage.

For Example:

docker run -p 8000:8000 kevaldekivadiya/textembed:latest --model sentence-transformers/all-MiniLM-L6-v2 --port 8000

Accessing the API

Once the server is running, you can access the API documentation via Swagger UI.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

textembed-0.0.3.tar.gz (19.7 kB view details)

Uploaded Source

Built Distribution

textembed-0.0.3-py3-none-any.whl (23.0 kB view details)

Uploaded Python 3

File details

Details for the file textembed-0.0.3.tar.gz.

File metadata

  • Download URL: textembed-0.0.3.tar.gz
  • Upload date:
  • Size: 19.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.9.19

File hashes

Hashes for textembed-0.0.3.tar.gz
Algorithm Hash digest
SHA256 80fefd7fae691cac783068e4f32b848bc005848ee672ca38bd7b744c8c91642f
MD5 4781a405a4f08725150e5fef24a97fe8
BLAKE2b-256 bffc966424317f8066ed1981b354c48220e04d9f3c71af1fb5dc4406e28d552b

See more details on using hashes here.

File details

Details for the file textembed-0.0.3-py3-none-any.whl.

File metadata

  • Download URL: textembed-0.0.3-py3-none-any.whl
  • Upload date:
  • Size: 23.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.9.19

File hashes

Hashes for textembed-0.0.3-py3-none-any.whl
Algorithm Hash digest
SHA256 0f43e995c0ae462265457f5f9ad1d2835a670bb0d41494f8b1cbe2b13a34dd8b
MD5 ba0af404a7b3f532af10f7cf4f2c9724
BLAKE2b-256 1ef4365fe64772960684d3f8633abdef4fedf11732159440143068a1a4b77696

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page