Skip to main content

Deploy DL/ ML inference pipelines with minimal extra code.

Project description

fastDeploy

Deploy DL/ ML inference pipelines with minimal extra code.

Installation:

pip install --upgrade fastdeploy

Usage:

# Invoke fastdeploy 
fastdeploy --help
# or
python -m fastdeploy --help

# Start prediction "loop" for recipe "deepsegment"
fastdeploy --recipe ./deepsegment --mode loop

# Start rest apis for recipe "deepsegment"
fastdeploy --recipe ./deepsegment --mode rest

# Run prediction using curl
curl -d '{"data": ["I was hungry i ordered a pizza"]}'\
-H "Content-Type: application/json" -X POST http://localhost:8080/infer

# Run prediction using python
python -c 'import requests; print(requests.post("http://localhost:8080/infer",\
json={"data": ["I was hungry i ordered a pizza"]}).json())'

# Response
{'prediction': [['I was hungry', 'i ordered a pizza']], 'success': True}

# Auto genereate dockerfile and build docker image. --base is docker base
fastdeploy --recipe ./recipes/deepsegment/ \
 --mode build_rest --base tensorflow/tensorflow:1.14.0-py3
# fastdeploy_deepsegment built!

# Run docker image
docker run -it -p8080:8080 fastdeploy_deepsegment

Features:

  1. Minimal extra code: No model exporting/ conversion/ freezing required. fastDeploy is the easiest way to serve and/or dockerize your existing inference code with minimal work.
  2. Fully configurable dynamic batching: fastDeploy dynamically batches concurrent requests for optimal resource usage.
  3. Containerization with no extra code: fastDeploy auto generates optimal dockerfiles and builds the image with no extra code.
  4. One consumer, multiple producers: (Coming soon) Single fastDeploy loop (consumer) can simultaneously be connected to multiple (types of) producers (rest, websocket, file).
  5. One producer, multiple consumers: Distribute one producer's work load to multiple consumers running on multiple nodes (assuming common storage is available for queues)

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

fastdeploy-1.0rc30.tar.gz (201.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

fastdeploy-1.0rc30-py2.py3-none-any.whl (13.3 kB view details)

Uploaded Python 2Python 3

File details

Details for the file fastdeploy-1.0rc30.tar.gz.

File metadata

  • Download URL: fastdeploy-1.0rc30.tar.gz
  • Upload date:
  • Size: 201.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.7.1 importlib_metadata/4.8.2 pkginfo/1.8.2 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.9.5

File hashes

Hashes for fastdeploy-1.0rc30.tar.gz
Algorithm Hash digest
SHA256 6afbbd6fdac8f05205f09406ebb6d27184d7c3b089d403c2684797ab4ab297ce
MD5 e44124931c0d06fb477cd73aa66b5706
BLAKE2b-256 89abceffd2b6f33157b2f2ea57562e83547408d254cfbf291bea22b2cef90245

See more details on using hashes here.

File details

Details for the file fastdeploy-1.0rc30-py2.py3-none-any.whl.

File metadata

  • Download URL: fastdeploy-1.0rc30-py2.py3-none-any.whl
  • Upload date:
  • Size: 13.3 kB
  • Tags: Python 2, Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.7.1 importlib_metadata/4.8.2 pkginfo/1.8.2 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.9.5

File hashes

Hashes for fastdeploy-1.0rc30-py2.py3-none-any.whl
Algorithm Hash digest
SHA256 99d4fbfd05046a14d54bf3b61a26ab24a344480b43af4791ca1535558ce79a33
MD5 65e97397dce0242b49b90818c4083c84
BLAKE2b-256 3317a8f1aeced27ef3c5cd165f4aad049a58835fadb85e612c0df2e9fe477b5b

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page