A high performance deep learning inference library
Project description
NVIDIA TensorRT is an SDK that facilitates high-performance machine learning inference. It is designed to work in a complementary fashion with training frameworks such as TensorFlow, PyTorch, and MXNet. It focuses specifically on running an already-trained network quickly and efficiently on NVIDIA hardware.
IMPORTANT: This is a special release of TensorRT designed to work only with TensorRT-LLM. Please refrain from upgrading to this version if you are not using TensorRT-LLM.
To install, please execute the following:
pip install tensorrt --extra-index-url https://pypi.nvidia.com
Or add the index URL to the (space-separated) PIP_EXTRA_INDEX_URL environment variable:
export PIP_EXTRA_INDEX_URL='https://pypi.nvidia.com'
pip install tensorrt
When the extra index url does not contain https://pypi.nvidia.com
, a nested pip install
will run with the proper extra index url hard-coded.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Hashes for tensorrt_dispatch-cu11-10.2.0.post1.tar.gz
Algorithm | Hash digest | |
---|---|---|
SHA256 | bf600aa0e70b06eeaa5cde9e9cb8a94a7a4bf5647dfcd1c044a1eb1939f25759 |
|
MD5 | 1f1aaac502b6f14081a17d5d399de849 |
|
BLAKE2b-256 | 88839927f290f4e9da6e0383db245c5eb516d5208b0638154c2df356eea488b8 |