Skip to main content

TextEmbed provides a robust and scalable REST API for generating vector embeddings from text. Built for performance and flexibility, it supports various sentence-transformer models, allowing users to easily integrate state-of-the-art NLP techniques into their applications. Whether you need embeddings for search, recommendation, or other NLP tasks, TextEmbed delivers with high efficiency.

Project description

TextEmbed - Embedding Inference Server

TextEmbed is a high-throughput, low-latency REST API designed for serving vector embeddings. It supports a wide range of sentence-transformer models and frameworks, making it suitable for various applications in natural language processing.

Features

  • High Throughput & Low Latency: Designed to handle a large number of requests efficiently.
  • Flexible Model Support: Works with various sentence-transformer models.
  • Scalable: Easily integrates into larger systems and scales with demand.
  • REST API: Simple and accessible API endpoints.

Getting Started

Prerequisites

Ensure you have Python 3.10 or higher installed. You will also need to install the required dependencies.

Installation

  1. Clone the repository:

    git clone https://github.com/kevaldekivadiya2415/textembed.git
    cd textembed
    
  2. Install the dependencies:

    pip install -r requirements.txt
    

Running the Server

  1. Set up the package in development mode:

    python3 setup.py develop
    
  2. Start the TextEmbed server with your desired model:

    python3 -m textembed.server --model <Model Name>
    

    Replace <Model Name> with the name of the model you want to use.

  3. For more information and additional options, run:

    python3 -m textembed.server --help
    

Running with Docker (Recommended)

You can also run TextEmbed using Docker. The Docker image is available on Docker Hub.

docker run keval2415/textembed:0.0.1 --help

This command will show the help message for the TextEmbed server, detailing the available options and usage.

For Example:

docker run -p 8000:8000 keval2415/textembed:0.0.1 --model sentence-transformers/all-MiniLM-L6-v2 --port 8000

Accessing the API

Once the server is running, you can access the API documentation via Swagger UI.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

textembed-0.0.2.tar.gz (18.8 kB view details)

Uploaded Source

Built Distribution

textembed-0.0.2-py3-none-any.whl (21.4 kB view details)

Uploaded Python 3

File details

Details for the file textembed-0.0.2.tar.gz.

File metadata

  • Download URL: textembed-0.0.2.tar.gz
  • Upload date:
  • Size: 18.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.10.13

File hashes

Hashes for textembed-0.0.2.tar.gz
Algorithm Hash digest
SHA256 c59a83433abbac0665b90dbd3c297a778c3791682dbd51e6e7bae661baedc747
MD5 9a9e456ff63751eb5cd4c0021ba6a3e9
BLAKE2b-256 106c5cce0cd8b79fd88a434dafac977758c1452dad06334b3a5171113e675fe0

See more details on using hashes here.

File details

Details for the file textembed-0.0.2-py3-none-any.whl.

File metadata

  • Download URL: textembed-0.0.2-py3-none-any.whl
  • Upload date:
  • Size: 21.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.10.13

File hashes

Hashes for textembed-0.0.2-py3-none-any.whl
Algorithm Hash digest
SHA256 e60fac8492fe7205830b5cc6266bcb081b8ed3839a5927f92a4f747c404092b7
MD5 9ed29e081bf8fd9844e438c68820b2f3
BLAKE2b-256 b3b748b074b2954c5c1606fcf0eab21ab6101fb42551e21304a34579fc838b01

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page