Skip to main content

TextEmbed provides a robust and scalable REST API for generating vector embeddings from text. Built for performance and flexibility, it supports various sentence-transformer models, allowing users to easily integrate state-of-the-art NLP techniques into their applications. Whether you need embeddings for search, recommendation, or other NLP tasks, TextEmbed delivers with high efficiency.

Project description

TextEmbed - Embedding Inference Server

TextEmbed is a high-throughput, low-latency REST API designed for serving vector embeddings. It supports a wide range of sentence-transformer models and frameworks, making it suitable for various applications in natural language processing.

Features

  • High Throughput & Low Latency: Designed to handle a large number of requests efficiently.
  • Flexible Model Support: Works with various sentence-transformer models.
  • Scalable: Easily integrates into larger systems and scales with demand.
  • Batch Processing: Supports batch processing for better and faster inference.
  • OpenAI Compatible REST API Endpoint: Provides an OpenAI compatible REST API endpoint.

Getting Started

Prerequisites

Ensure you have Python 3.11 or higher installed. You will also need to install the required dependencies.

Installation

  1. Install the required dependencies:

    pip install -U textembed
    
  2. Start the TextEmbed server with your desired model:

    python3 -m textembed.server --model <Model Name>
    

    Replace <Model Name> with the name of the model you want to use.

  3. For more information and additional options, run:

    python3 -m textembed.server --help
    

Running with Docker (Recommended)

You can also run TextEmbed using Docker. The Docker image is available on Docker Hub.

docker run kevaldekivadiya/textembed:latest --help

This command will show the help message for the TextEmbed server, detailing the available options and usage.

For Example:

docker run -p 8000:8000 kevaldekivadiya/textembed:latest --model sentence-transformers/all-MiniLM-L6-v2 --port 8000

Accessing the API

Once the server is running, you can access the API documentation via Swagger UI.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

textembed-0.0.5.tar.gz (20.1 kB view details)

Uploaded Source

Built Distribution

textembed-0.0.5-py3-none-any.whl (23.4 kB view details)

Uploaded Python 3

File details

Details for the file textembed-0.0.5.tar.gz.

File metadata

  • Download URL: textembed-0.0.5.tar.gz
  • Upload date:
  • Size: 20.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.9.19

File hashes

Hashes for textembed-0.0.5.tar.gz
Algorithm Hash digest
SHA256 c820bbc6dd57511c7627caafff9764ff71e36cd3e8ba10174f16b62d0aab1fb8
MD5 84f7f3bf445f22f9ea9423d71ff5ef9d
BLAKE2b-256 d2a3f3b27f5cdd7e715a14ea29f04c46544b3dd5b23b6686c4fbfb3a37fd9cf3

See more details on using hashes here.

File details

Details for the file textembed-0.0.5-py3-none-any.whl.

File metadata

  • Download URL: textembed-0.0.5-py3-none-any.whl
  • Upload date:
  • Size: 23.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.9.19

File hashes

Hashes for textembed-0.0.5-py3-none-any.whl
Algorithm Hash digest
SHA256 2c85e5537f719dba2f566368307a24869af7d0bd3c1574d62e2a44a9d1cf7b6f
MD5 984f0ad66ebf56d5e226786239922e35
BLAKE2b-256 3e870d1d56b087c4a15b4ce5cd7700befa83a591f5f45350c6922425890ab360

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page