TextEmbed provides a robust and scalable REST API for generating vector embeddings from text. Built for performance and flexibility, it supports various sentence-transformer models, allowing users to easily integrate state-of-the-art NLP techniques into their applications. Whether you need embeddings for search, recommendation, or other NLP tasks, TextEmbed delivers with high efficiency.
Project description
TextEmbed - Embedding Inference Server
TextEmbed is a high-throughput, low-latency REST API designed for serving vector embeddings. It supports a wide range of sentence-transformer models and frameworks, making it suitable for various applications in natural language processing.
Features
- High Throughput & Low Latency: Designed to handle a large number of requests efficiently.
- Flexible Model Support: Works with various sentence-transformer models.
- Scalable: Easily integrates into larger systems and scales with demand.
- Batch Processing: Supports batch processing for better and faster inference.
- OpenAI Compatible REST API Endpoint: Provides an OpenAI compatible REST API endpoint.
Getting Started
Prerequisites
Ensure you have Python 3.11 or higher installed. You will also need to install the required dependencies.
Installation
-
Install the required dependencies:
pip install -r requirements.txt
-
Start the TextEmbed server with your desired model:
python3 -m textembed.server --model <Model Name>
Replace
<Model Name>
with the name of the model you want to use. -
For more information and additional options, run:
python3 -m textembed.server --help
Running with Docker (Recommended)
You can also run TextEmbed using Docker. The Docker image is available on Docker Hub.
docker run kevaldekivadiya/textembed:latest --help
This command will show the help message for the TextEmbed server, detailing the available options and usage.
For Example:
docker run -p 8000:8000 kevaldekivadiya/textembed:latest --model sentence-transformers/all-MiniLM-L6-v2 --port 8000
Accessing the API
Once the server is running, you can access the API documentation via Swagger UI.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file textembed-0.0.3.tar.gz
.
File metadata
- Download URL: textembed-0.0.3.tar.gz
- Upload date:
- Size: 19.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.0 CPython/3.9.19
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 80fefd7fae691cac783068e4f32b848bc005848ee672ca38bd7b744c8c91642f |
|
MD5 | 4781a405a4f08725150e5fef24a97fe8 |
|
BLAKE2b-256 | bffc966424317f8066ed1981b354c48220e04d9f3c71af1fb5dc4406e28d552b |
File details
Details for the file textembed-0.0.3-py3-none-any.whl
.
File metadata
- Download URL: textembed-0.0.3-py3-none-any.whl
- Upload date:
- Size: 23.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.0 CPython/3.9.19
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 0f43e995c0ae462265457f5f9ad1d2835a670bb0d41494f8b1cbe2b13a34dd8b |
|
MD5 | ba0af404a7b3f532af10f7cf4f2c9724 |
|
BLAKE2b-256 | 1ef4365fe64772960684d3f8633abdef4fedf11732159440143068a1a4b77696 |