Skip to main content

Deploy DL/ ML inference pipelines with minimal extra code.

Project description

fastDeploy

Deploy DL/ ML inference pipelines with minimal extra code.

Installation:

pip install --upgrade fastdeploy

Usage:

# Invoke fastdeploy 
fastdeploy --help
# or
python -m fastdeploy --help

# Start prediction "loop" for recipe "echo_json"
fastdeploy --recipe ./echo_json --mode loop

# Start rest apis for recipe "echo_json"
fastdeploy --recipe ./echo_json --mode rest

# Auto genereate dockerfile and build docker image. --base is docker base
fastdeploy --recipe ./recipes/echo_json/ \
 --mode build_rest --base python:3.6-slim
# fastdeploy_echo_json built!

# Run docker image
docker run -it -p8080:8080 fastdeploy_echo_json

fastDeploy monitor

  • available on localhost:8080 (or --port)

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

fastdeploy-2.2.15.tar.gz (195.6 kB view hashes)

Uploaded Source

Built Distribution

fastdeploy-2.2.15-py3-none-any.whl (160.3 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page