Skip to main content

A python library and framework for fast neural network computations.

Project description

Fast Neural Networks (FastNN)

A framework for deploying serializable and optimizable neural net models at scale in production via. the NVIDIA Triton Inference Server.

FastNN Docker Release Selector (Ubuntu 18.04)

Documentation - Guides, Models, API References

Features:

  • Data Processing
    • Intuitive data processing modules for encoding human-readible data into tensors compatible with deep learning libraries
  • Model Exporting
    • FastNN torch modules and tools for exporting models via. TorchScript tracing and scripting to a production environment. Now includes Text Generation.
  • Model Zoo
    • Various exported models hosted in this repo via. git-lfs and AWS S3. Includes models from the HuggingFace's Transformers and TorchVision
  • Model Deployment
    • Deploy models using Triton Inference Server on CPU/GPU-compatible server(s) with helm or docker
  • FastNN Client
    • Client wrapper for Triton Inference Server's client module for programmatic requests with python

Pre-Requisites:

Git LFS is required if you'd like to use any of the models provided by FastNN in ./model_repository.

Cloning this repository without Git LFS will clone a repository with LFS pointers, not the actual model.

After the repository is cloned and Git LFS is installed, set up with git lfs install.

Download specific models with:

git lfs pull --include="model_repository/<path-to-model-dir>" --exclude=""

Download ALL models with:

git lfs pull

# Or
#git lfs pull --include="*" --exclude=""

Quickstart and Installation:

Pre-requisites:

Requirements: Python 3.9+, PyTorch 2+, Triton Client

Optional: CUDA Compatible GPU, NVIDIA Drivers, cudnn (PyTorch pre-built wheels)

  1. To install PyTorch with TorchVision, please refer to the installation instructions on their web page here.

  2. The tritonclient package wheels are not hosted on the public PyPI server. We need to add the address of NVIDA's private python package index to the environment. You can complete these steps and install the tritonclient package by running the following.

# If you cloned this repo, you can just uncomment and run the one line below
#sh ./scripts/install_triton_client.
pip install nvidia-pyindex
pip install tritonclient[all]

Install via. pip

Once the above requirements are set, you can install fastnn with the command below:

pip install fastnn

If you are comfortable with the latest default stable releases of PyTorch you can skip step 1 in the pre-requisites and run pip install fastnn[torch] instead.

Install from source with Poetry for development

You will need to install poetry by referring to the installation instructions on their web page here.

After cloning the repository, just run poetry install to install with the .lock file.

Activate the virtual environment with the command below:

poetry shell

Docker

Official FastNN images are hosted on Docker Hub.

Select FastNN package and image versions by referencing the documentation. Development and runtime environments are available.

Jupyter lab and notebook servers are accessible with notebook examples and terminal access localhost:8888 with every image.

Run the latest FastNN image by running below:

docker run --gpus all --rm -it -p 8888:8888 aychang/fastnn:latest

Run images with specific configurations as can see in the example command below:

docker run --gpus all --rm -it -p 8888:8888 aychang/fastnn:0.1.0-cuda11.0-runtime-ubuntu18.04-py3.7

Triton Inference Server: Local or Kubernetes Cluster

Local Deployment

Requirements:

  • Docker 19.03+

GPU Inference Requirements:

Local deployment of the Triton Server uses the EXPLICIT model control mode. Local models must be explicitly specified with the --load-model argument in ./scripts/start_triton_local.sh

export MODEL_REPOSITORY=$(pwd)/model_repository
sh ./scripts/start_triton_local.sh

Helm Chart install in Kubernetes Cluster

Requirements: kubectl 1.17+, Helm 3+, Kubernetes 1.17+

You can currently install the local FastNN helm chart with the following instuctions:

cd ./k8s
helm install fastnn .
export MODEL_REPOSITORY=$(pwd)/model_repository

Note: The current local helm chart installation deploys Triton using the NONE model control mode. All models available in the S3 Model Zoo will be deployed...good luck. Deployed models and model control mode can be edited in the helm chart deployment configuration file.

License

This project is licensed under the terms of the MIT license.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

fastnn-0.4.0.tar.gz (28.8 kB view details)

Uploaded Source

Built Distribution

fastnn-0.4.0-py3-none-any.whl (33.3 kB view details)

Uploaded Python 3

File details

Details for the file fastnn-0.4.0.tar.gz.

File metadata

  • Download URL: fastnn-0.4.0.tar.gz
  • Upload date:
  • Size: 28.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.6.1 CPython/3.9.18 Linux/6.2.0-1015-azure

File hashes

Hashes for fastnn-0.4.0.tar.gz
Algorithm Hash digest
SHA256 93b01b7dcb8ea22f131eda9cad22bdef555e78d8dde58b5fcd49df87411f7f70
MD5 2c5de5794f1f56ec5ab88a10cf488fb9
BLAKE2b-256 a92d7f964c3926d8f65fc8f82c1a7ee12528d9238eded2588cb19c68cfd9637e

See more details on using hashes here.

File details

Details for the file fastnn-0.4.0-py3-none-any.whl.

File metadata

  • Download URL: fastnn-0.4.0-py3-none-any.whl
  • Upload date:
  • Size: 33.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.6.1 CPython/3.9.18 Linux/6.2.0-1015-azure

File hashes

Hashes for fastnn-0.4.0-py3-none-any.whl
Algorithm Hash digest
SHA256 d1574665559b74a19a0dec850bbdc8141c8bd49faee3c4e8da6f70e1a3d33755
MD5 d22a4e2f91889f7cfde44dea9c31eaa6
BLAKE2b-256 87163e4eebebbf440cf8977077f845b807f271e920e1b549dc9acf12b4c95dd5

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page