Skip to main content

Visualization of Pytorch Models

Project description

torchview

Python 3.7+ PyPI version Conda version Build Status GitHub license codecov Downloads

Torchview provides visualization of pytorch models in the form of visual graphs. Visualization includes tensors, modules, torch.functions and info such as input/output shapes.

Pytorch version of plot_model of keras

Supports PyTorch versions $\geq$ 1.7.

Installation

First, you need to install graphviz,

pip install graphviz

For python interface of graphiz to work, you need to have dot layout command working in your system. If it isn't already installed, I suggest you run the following depeding on your OS,

Debian-based Linux distro (e.g. Ubuntu):

apt-get install graphviz

Windows:

choco install graphviz

macOS:

brew install graphviz

see more details here

Then, continue with installing torchview using pip

pip install torchview

or if you want via conda

conda install -c conda-forge torchview

How To Use

from torchview import draw_graph

model = MLP()
batch_size = 2
model_graph = draw_graph(model, input_size=(batch_size, 128))
model_graph.visual_graph

output

Notebook Examples

For more examples, see colab notebooks below,

Introduction Notebook: Introduction

Computer Vision Models: Vision

NLP Models: NLP

Note: Output graphviz visuals return images with desired sizes. But sometimes, on VSCode, some shapes are being cropped due to large size and svg rendering on by VSCode. To solve this, I suggest you run the following

import graphviz
graphviz.set_jupyter_format('png')

This problem does not occur on other platforms e.g. JupyterLab or Google Colab.

Supported Features

  • Almost all the models, RNN, Sequentials, Skip Connection
  • Shows operations between tensors (in addition to module calls)
  • Rolling/Unrolling feature. Recursively used modules can be rolled visually, see below.

Documentation

def draw_graph(
    model: nn.Module,
    input_data: INPUT_DATA_TYPE | None = None,
    input_size: INPUT_SIZE_TYPE | None = None,
    graph_name: str = 'model',
    depth: int | float = 3,
    device: torch.device | str | None = None,
    dtypes: list[torch.dtype] | None = None,
    mode: str | None = None,
    strict: bool = True,
    expand_nested: bool = False,
    hide_module_functions: bool = True,
    hide_inner_tensors: bool = True,
    roll: bool = False,
    show_shapes: bool = True,
    save_graph: bool = False,
    filename: str | None = None,
    directory: str = '.',
    **kwargs: Any,
) -> ComputationGraph:
    '''Returns visual representation of the input Pytorch Module with
    ComputationGraph object. ComputationGraph object contains:

    1) Root nodes (usually tensor node for input tensors) which connect to all
    the other nodes of computation graph of pytorch module recorded during forward
    propagation.

    2) graphviz.Digraph object that contains visual representation of computation
    graph of pytorch module. This graph visual shows modules/ module hierarchy,
    torch_functions, shapes and tensors recorded during forward prop, for examples
    see documentation, and colab notebooks.


    Args:
        model (nn.Module):
            Pytorch model to represent visually.

        input_data (data structure containing torch.Tensor):
            input for forward method of model. Wrap it in a list for
            multiple args or in a dict or kwargs

        input_size (Sequence of Sizes):
            Shape of input data as a List/Tuple/torch.Size
            (dtypes must match model input, default is FloatTensors).
            Default: None

        graph_name (str):
            Name for graphviz.Digraph object. Also default name graphviz file
            of Graph Visualization
            Default: 'model'

        depth (int):
            Upper limit for depth of nodes to be shown in visualization.
            Depth is measured how far is module/tensor inside the module hierarchy.
            For instance, main module has depth=0, whereas submodule of main module
            has depth=1, and so on.
            Default: 3

        device (str or torch.device):
            Device to place and input tensors. Defaults to
            gpu if cuda is seen by pytorch, otherwise to cpu.
            Default: None

        dtypes (list of torch.dtype):
            Uses dtypes to set the types of input tensor if
            input size is given.

        mode (str):
            Mode of model to use for forward prop. Defaults
            to Eval mode if not given
            Default: None

        strict (bool):
            if true, graphviz visual does not allow multiple edges
            between nodes. Mutiple edge occurs e.g. when there are tensors
            from module node to module node and hiding those tensors
            Default: True
        
        expand_nested(bool):
            if true shows nested modules with dashed borders

        hide_module_function (bool):
            Determines whether to hide module torch_functions. Some
            modules consist only of torch_functions (no submodule),
            e.g. nn.Conv2d.
            True => Dont include module functions in graphviz
            False => Include modules function in graphviz
            Default: True

        hide_inner_tensors (bool):
            Inner tensor is all the tensors of computation graph
            but input and output tensors
            True => Does not show inner tensors in graphviz
            False => Shows inner tensors in graphviz
            Default: True

        roll (bool):
            If true, rolls recursive modules.
            Default: False

        show_shapes (bool):
            True => Show shape of tensor, input, and output
            False => Dont show
            Default: True

        save_graph (bool):
            True => Saves output file of graphviz graph
            False => Does not save
            Default: False

        filename (str):
            name of the file to store dot syntax representation and
            image file of graphviz graph. Defaults to graph_name

        directory (str):
            directory in which to store graphviz output files.
            Default: .

    Returns:
        ComputationGraph object that contains visualization of the input
        pytorch model in the form of graphviz Digraph object
    '''

Examples

Rolled Version of Recursive Networks

from torchview import draw_graph

model_graph = draw_graph(
    SimpleRNN(), input_size=(2, 3),
    graph_name='RecursiveNet',
    roll=True
)
model_graph.visual_graph

rnns

Show/Hide intermediate (hidden) tensors and Functionals

# Show inner tensors and Functionals
model_graph = draw_graph(
    MLP(), input_size=(2, 128),
    graph_name='MLP',
    hide_inner_tensors=False,
    hide_module_functions=False,
)

model_graph.visual_graph

download

ResNet / Skip Connection / Support for Torch operations / Nested Modules

import torchvision

model_graph = draw_graph(resnet18(), input_size=(1,3,32,32), expand_nested=True)
model_graph.visual_graph

expand_nested_resnet_model gv

TODO:

  • Display Module parameter info
  • Support for Graph Neural Network (GNN)
  • Support for Undirected edges for GNNs
  • Support for torch-based functions[^1]

[^1]: Here, torch-based functions refers to any function that uses only torch functions and modules. This is more general than modules.

Contributing

All issues and pull requests are much appreciated! If you are wondering how to build the project:

  • torchview is actively developed using the latest version of Python.
    • Changes should be backward compatible to Python 3.7, and will follow Python's End-of-Life guidance for old versions.
    • Run pip install -r requirements-dev.txt. We use the latest versions of all dev packages.
    • To run unit tests, run pytest.
    • To update the expected output files, run pytest --overwrite.
    • To skip output file tests, use pytest --no-output

References

  • Parts related to input processing and validation are taken/inspired from torchinfo repository!!.
  • Many of the software related parts (e.g. CI, testing) are also taken/inspired from torchinfo repository. So big thanks to @TylerYep!!!
  • The algorithm for constructing visual graphs is thanks to __torch_function__ and subclassing torch.Tensor. Big thanks to all those who developed this API!!.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

torchview-0.2.2.tar.gz (24.3 kB view details)

Uploaded Source

Built Distribution

torchview-0.2.2-py3-none-any.whl (24.3 kB view details)

Uploaded Python 3

File details

Details for the file torchview-0.2.2.tar.gz.

File metadata

  • Download URL: torchview-0.2.2.tar.gz
  • Upload date:
  • Size: 24.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.10.6

File hashes

Hashes for torchview-0.2.2.tar.gz
Algorithm Hash digest
SHA256 fee4c18d88728acf954cbaab2bb478d935762eea63a234ffd0bba9da5abd8467
MD5 fb8569facb35864112d49443d4d9a81b
BLAKE2b-256 2b6b1a4ce5cb1033e3c4379b9534f66bf9c3b156e45d65b24aa8a69ac05718b6

See more details on using hashes here.

File details

Details for the file torchview-0.2.2-py3-none-any.whl.

File metadata

  • Download URL: torchview-0.2.2-py3-none-any.whl
  • Upload date:
  • Size: 24.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.10.6

File hashes

Hashes for torchview-0.2.2-py3-none-any.whl
Algorithm Hash digest
SHA256 d3233116e413bc436b3dd97bdb82bf37af4ee611f2ca599d40c154311b669a0a
MD5 27b9e04d81ce88295ee72b1f556f4438
BLAKE2b-256 641606c2deb040b3f72e637b328e954afdf409a122d9c14b1c247b9de8f184fb

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page