Skip to main content

Lightning Thunder project.

Project description

Thunder Thunder

Make PyTorch models Lightning fast.


Lightning.aiPerformanceGet startedInstallExamplesInside ThunderGet involved!Documentation

license CI testing General checks Documentation Status pre-commit.ci status

Welcome to ⚡ Lightning Thunder

Thunder makes PyTorch models Lightning fast.

Thunder is a source-to-source compiler for PyTorch. It makes PyTorch programs faster by combining and using different hardware executors at once (for instance, nvFuser, torch.compile, cuDNN, and TransformerEngine FP8).

It supports both single and multi-GPU configurations. Thunder aims to be usable, understandable, and extensible.

 

[!Note] Lightning Thunder is in alpha. Feel free to get involved, but expect a few bumps along the way.

 

Single-GPU performance

Thunder can achieve significant speedups over standard non-compiled PyTorch code ("PyTorch eager"), through the compounding effects of optimizations and the use of best-in-class executors. The figure below shows the pretraining throughput for Llama 2 7B as implemented in LitGPT.

Thunder

As shown in the plot above, Thunder achieves a 40% speedup in training throughput compared to eager code on H100 using a combination of executors including nvFuser, torch.compile, cuDNN, and TransformerEngine FP8.

 

Multi-GPU performance

Thunder also supports distributed strategies such as DDP and FSDP for training models on multiple GPUs. The following plot displays the normalized throughput measured for Llama 2 7B without FP8 mixed precision; support for FSDP is in progress.

Thunder

 

Get started

The easiest way to get started with Thunder, requiring no extra installations or setups, is by using our Zero to Thunder Tutorial Studio.

 

Install Thunder

To use Thunder on your local machine:

  • install nvFuser nightly and PyTorch nightly together as follows:
# install nvFuser which installs the matching nightly PyTorch
pip install --pre 'nvfuser-cu121[torch]' --extra-index-url https://pypi.nvidia.com
  • install cudnn as follows:
# install cudnn
pip install nvidia-cudnn-frontend
  • Finally, install Thunder as follows:
# install thunder
pip install lightning-thunder
Advanced install options

 

Install from main

Alternatively, you can install the latest version of Thunder directly from this GitHub repository as follows:

# 1) Install nvFuser and PyTorch nightly dependencies:
pip install --pre 'nvfuser-cu121[torch]' --extra-index-url https://pypi.nvidia.com
# 2) Install Thunder itself
pip install git+https://github.com/Lightning-AI/lightning-thunder.git

 

Install to tinker and contribute

If you are interested in tinkering with and contributing to Thunder, we recommend cloning the Thunder repository and installing it in pip's editable mode:

git clone https://github.com/Lightning-AI/lightning-thunder.git
cd lightning-thunder
pip install -e .

 

Develop and run tests

After cloning the lightning-thunder repository and installing it as an editable package as explained above, ou can set up your environment for developing Thunder by installing the development requirements:

pip install -r requirements/devel.txt

Now you run tests:

pytest thunder/tests

Thunder is very thoroughly tested, so expect this to take a while.

 

Hello World

Below is a simple example of how Thunder allows you to compile and run PyTorch code:

import torch
import thunder


def foo(a, b):
    return a + b


jfoo = thunder.jit(foo)

a = torch.full((2, 2), 1)
b = torch.full((2, 2), 3)

result = jfoo(a, b)

print(result)

# prints
# tensor(
#  [[4, 4]
#   [4, 4]])

The compiled function jfoo takes and returns PyTorch tensors, just like the original function, so modules and functions compiled by Thunder can be used as part of larger PyTorch programs.

 

Train models

Thunder is in its early stages and should not be used for production runs yet.

However, it can already deliver outstanding performance for pretraining and finetuning LLMs supported by LitGPT, such as Mistral, Llama 2, Gemma, Falcon, and others.

Check out the LitGPT integration to learn about running LitGPT and Thunder together.

 

Inside Thunder: A brief look at the core features

Given a Python callable or PyTorch module, Thunder can generate an optimized program that:

  • Computes its forward and backward passes
  • Coalesces operations into efficient fusion regions
  • Dispatches computations to optimized kernels
  • Distributes computations optimally across machines

To do so, Thunder ships with:

  • A JIT for acquiring Python programs targeting PyTorch and custom operations
  • A multi-level intermediate representation (IR) to represent operations as a trace of a reduced operation set
  • An extensible set of transformations on the trace of a computational graph, such as grad, fusions, distributed (like ddp, fsdp), functional (like vmap, vjp, jvp)
  • A way to dispatch operations to an extensible collection of executors

Thunder is written entirely in Python. Even its trace is represented as valid Python at all stages of transformation. This allows unprecedented levels of introspection and extensibility.

Thunder doesn't generate code for accelerators, such as GPUs, directly. It acquires and transforms user programs so that it's possible to optimally select or generate device code using fast executors like:

Modules and functions compiled with Thunder fully interoperate with vanilla PyTorch and support PyTorch's autograd. Also, Thunder works alongside torch.compile to leverage its state-of-the-art optimizations.

 

Documentation

Online documentation is available. To build documentation locally you can use

make docs

and point your browser to the generated docs at docs/build/index.html.

 

Get involved!

We appreciate your feedback and contributions. If you have feature requests, questions, or want to contribute code or config files, please don't hesitate to use the GitHub Issue tracker.

We welcome all individual contributors, regardless of their level of experience or hardware. Your contributions are valuable, and we are excited to see what you can accomplish in this collaborative and supportive environment.

 

License

Lightning Thunder is released under the Apache 2.0 license. See the LICENSE file for details.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lightning_thunder-0.2.0.dev20240728.tar.gz (458.7 kB view details)

Uploaded Source

Built Distribution

File details

Details for the file lightning_thunder-0.2.0.dev20240728.tar.gz.

File metadata

File hashes

Hashes for lightning_thunder-0.2.0.dev20240728.tar.gz
Algorithm Hash digest
SHA256 58c6fa0f491edbcb22b50ad01b592e08320c2474c2319f2bf5d787f0ff37191a
MD5 2655744019b810a28664a84ded8ca1e6
BLAKE2b-256 62c36fcb3da504d76c45c842cdfd59dbc541409de1759cc902e17ce007ac7c39

See more details on using hashes here.

File details

Details for the file lightning_thunder-0.2.0.dev20240728-py3-none-any.whl.

File metadata

File hashes

Hashes for lightning_thunder-0.2.0.dev20240728-py3-none-any.whl
Algorithm Hash digest
SHA256 004a3528c7e46bdd3e7c9816f4080511f2469df9ee5f8444c21bb41845141c15
MD5 b530c51437349d8f860f145d7b33a7d6
BLAKE2b-256 c0eaa20bc132ae50d4c278697df098186a9197fa65962a8d9856444e33c5646f

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page