Deploy DL/ ML inference pipelines with minimal extra code.
Project description
fastDeploy
Deploy DL/ ML inference pipelines with minimal extra code.
Installation:
pip install --upgrade fastdeploy
Usage:
# Invoke fastdeploy
fastdeploy --help
# or
python -m fastdeploy --help
# Start prediction "loop" for recipe "deepsegment"
fastdeploy --recipe ./deepsegment --mode loop
# Start rest apis for recipe "deepsegment"
fastdeploy --recipe ./deepsegment --mode rest
# Run prediction using curl
curl -d '{"data": ["I was hungry i ordered a pizza"]}'\
-H "Content-Type: application/json" -X POST http://localhost:8080/infer
# Run prediction using python
python -c 'import requests; print(requests.post("http://localhost:8080/infer",\
json={"data": ["I was hungry i ordered a pizza"]}).json())'
# Response
[{'prediction': [['I was hungry', 'i ordered a pizza']], 'success': True}, '200 OK']
# Auto genereate dockerfile and build docker image. --base is docker base
fastdeploy --recipe ./recipes/deepsegment/ \
--mode build_rest --base tensorflow/tensorflow:1.14.0-py3
# fastdeploy_deepsegment built!
# Run docker image
docker run -it -p8080:8080 fastdeploy_deepsegment
# pip version of fastdeploy uses pywsgi with gevent.
# Docker version uses by default 3 gunicorn gevent workers with 1000 "threads" per worker.
Features:
- Minimal extra code: No model exporting/ conversion/ freezing required. fastDeploy is the easiest way to serve and/or dockerize your existing inference code with minimal work.
- Fully configurable dynamic batching: fastDeploy dynamically batches concurrent requests for optimal resource usage.
- Containerization with no extra code: fastDeploy auto generates optimal dockerfiles and builds the image with no extra code.
- One consumer, multiple producers: Single fastDeploy loop (consumer) can simultaneously be connected to multiple (types of) producers (rest, websocket, file).
- One producer, multiple consumers: Distribute one producer's work load to multiple consumers running on multiple nodes (assuming common storage is available for queues)
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
fastdeploy-1.0rc20.tar.gz
(11.4 kB
view hashes)
Built Distribution
Close
Hashes for fastdeploy-1.0rc20-py2.py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 46fff8ea84361fc7996aff45a08a7f61105b43a781401270adfce15b2679349c |
|
MD5 | 74d4cfa7f2b10366f3fae1594e806d23 |
|
BLAKE2b-256 | b705730a34a4f7e79c200c57ddd33d5efa7ec13f2bb3e730cf7ff613d41a9551 |