Deploy DL/ ML inference pipelines with minimal extra code.
Project description
fastDeploy
Deploy DL/ ML inference pipelines with minimal extra code.
Installation:
pip install --upgrade fastdeploy
Usage:
# Invoke fastdeploy
fastdeploy --help
# or
python -m fastdeploy --help
# Start prediction "loop" for recipe "deepsegment"
fastdeploy --recipe ./deepsegment --mode loop
# Start rest apis for recipe "deepsegment"
fastdeploy --recipe ./deepsegment --mode rest
# Run json prediction using curl
curl -d '{"data": ["I was hungry i ordered a pizza"]}'\
-H "Content-Type: application/json" -X POST http://localhost:8080/infer
# Run file prediction using curl
curl -F image_1=@image_1.png image_2=@image_2.png http://localhost:8080/infer
# Run file prediction using python
python -c 'import requests; print(requests.post("http://localhost:8080/infer",\
json={"data": ["I was hungry i ordered a pizza"]}).json())'
# Run prediction using python
python -c 'import requests; print(requests.post("http://localhost:8080/infer",\
json={"data": ["I was hungry i ordered a pizza"]}).json())'
# Response
{'prediction': [['I was hungry', 'i ordered a pizza']], 'success': True}
# Auto genereate dockerfile and build docker image. --base is docker base
fastdeploy --recipe ./recipes/deepsegment/ \
--mode build_rest --base tensorflow/tensorflow:1.14.0-py3
# fastdeploy_deepsegment built!
# Run docker image
docker run -it -p8080:8080 fastdeploy_deepsegment
Features:
- Minimal extra code: No model exporting/ conversion/ freezing required. fastDeploy is the easiest way to serve and/or dockerize your existing inference code with minimal work.
- Fully configurable dynamic batching: fastDeploy dynamically batches concurrent requests for optimal resource usage.
- Containerization with no extra code: fastDeploy auto generates optimal dockerfiles and builds the image with no extra code.
- One consumer, multiple producers: (Coming soon) Single fastDeploy loop (consumer) can simultaneously be connected to multiple (types of) producers (rest, websocket, file).
- One producer, multiple consumers: Distribute one producer's work load to multiple consumers running on multiple nodes (assuming common storage is available for queues)
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
fastdeploy-1.0rc36.tar.gz
(12.4 kB
view details)
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file fastdeploy-1.0rc36.tar.gz.
File metadata
- Download URL: fastdeploy-1.0rc36.tar.gz
- Upload date:
- Size: 12.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.8.0 pkginfo/1.8.2 readme-renderer/33.0 requests/2.27.1 requests-toolbelt/0.9.1 urllib3/1.26.8 tqdm/4.63.0 importlib-metadata/4.11.2 keyring/23.5.0 rfc3986/2.0.0 colorama/0.4.4 CPython/3.10.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
05e4e562a7a6654b41f720e2c363241235cda4e4c5864bbb6a99ee4e3f79dd3f
|
|
| MD5 |
9a7b9860e2387015492a0a499baea26b
|
|
| BLAKE2b-256 |
88323838b3386bf5ff77c25f7192ae3cf855153280efceeff10f1c7113ff854d
|
File details
Details for the file fastdeploy-1.0rc36-py3-none-any.whl.
File metadata
- Download URL: fastdeploy-1.0rc36-py3-none-any.whl
- Upload date:
- Size: 13.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.8.0 pkginfo/1.8.2 readme-renderer/33.0 requests/2.27.1 requests-toolbelt/0.9.1 urllib3/1.26.8 tqdm/4.63.0 importlib-metadata/4.11.2 keyring/23.5.0 rfc3986/2.0.0 colorama/0.4.4 CPython/3.10.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
2a4d72ce82bda0b9e2234d3b59683d4317b1e26fa6d37566a47c264c5f756eb1
|
|
| MD5 |
a6e1f3a3becd1e132c62afe729f157bd
|
|
| BLAKE2b-256 |
31d83ed389c3a8f17f3acfba1deccaef2f54ea3126223a44df764f8bd75a50cd
|