Skip to main content

Tripy: A Python Programming Model For TensorRT

Project description

Tripy: A Python Programming Model For TensorRT

Quick Start | Installation | Examples | Notebooks | Contributing | Documentation

Tripy L1

Tripy is a debuggable, Pythonic frontend for TensorRT, a deep learning inference compiler.

What you can expect:

  • High performance by leveraging TensorRT's optimization capabilties.
  • An intuitive API that follows conventions of the ecosystem.
  • Debuggability with features like eager mode to interactively debug mistakes.
  • Excellent error messages that are informative and actionable.
  • Friendly documentation that is comprehensive but concise, with code examples.

Installation

Option 1: Install with pip

python3 -m pip install nvtripy -f https://nvidia.github.io/TensorRT-Incubator/packages.html

Option 2: Use the container image

docker run --pull always --gpus all -it --rm ghcr.io/nvidia/tensorrt-incubator/nvtripy:latest

Quick Start

See the Introduction To Tripy guide for details:

  • Defining a model:

    class Model(tp.Module):
        def __init__(self):
            self.conv = tp.Conv(in_channels=1, out_channels=1, kernel_dims=[3, 3])
    
        def forward(self, x):
            x = self.conv(x)
            x = tp.relu(x)
            return x
    
  • Initializing it:

    model = Model()
    model.load_state_dict(
        {
            "conv.weight": tp.ones((1, 1, 3, 3)),
            "conv.bias": tp.ones((1,)),
        }
    )
    
    dummy_input = tp.ones((1, 1, 4, 4)).eval()
    
  • Executing in eager mode:

    eager_out = model(dummy_input)
    
  • Compiling and executing:

    compiled_model = tp.compile(
        model,
        args=[tp.InputInfo(shape=(1, 1, 4, 4), dtype=tp.float32)],
    )
    
    compiled_out = compiled_model(dummy_input)
    

Building Wheels From Source

For the latest changes, build Tripy wheels from source:

  1. Install build:

    python3 -m pip install build
    
  2. Build a wheel from the tripy root directory:

    python3 -m build . -w
    
  3. Install the wheel from the tripy root directory:

    python3 -m pip install -f https://nvidia.github.io/TensorRT-Incubator/packages.html dist/nvtripy-*.whl
    
  4. [Optional] Sanity check:

    python3 -c "import nvtripy as tp; x = tp.ones((5,), dtype=tp.int32); assert x.tolist() == [1] * 5"
    

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

nvtripy-0.1.7-py3-none-any.whl (294.5 kB view details)

Uploaded Python 3

File details

Details for the file nvtripy-0.1.7-py3-none-any.whl.

File metadata

  • Download URL: nvtripy-0.1.7-py3-none-any.whl
  • Upload date:
  • Size: 294.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for nvtripy-0.1.7-py3-none-any.whl
Algorithm Hash digest
SHA256 3a9b7e6f64c6c8b36852d282312fc71205522ab34bba5c1d9f48afdd7a077a70
MD5 9f570dc9b4f60dadf55e91935854ee76
BLAKE2b-256 1f491a5abe6cc057cbe3f0c4a0e4b542b9cba3bbacb544fbc1642a9661965c80

See more details on using hashes here.

Provenance

The following attestation bundles were made for nvtripy-0.1.7-py3-none-any.whl:

Publisher: tripy-release.yml on NVIDIA/TensorRT-Incubator

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page