Skip to main content

Simplification of pruned models for accelerated inference

Project description

Simplify

Simplification of pruned models for accelerated inference.

Fully supports torchvision classification models up to version 0.11.

Installation

Simplify can be installed using pip:

pip3 install torch-simplify

or if you want to run the latest version of the code, you can install from git:

git clone https://github.com/EIDOSlab/simplify
cd simplify
pip3 install -r requirements.txt

Usage

Main function

For most scenarios the main simplify function will suffice. This function returns the simplified model.

Arguments

The expected arguments are:

  • model (torch.nn.Module): Module to be simplified i.e. the PyTorch's model.
  • x (torch.Tensor): zero-tensor of shape [1, C, N, M], same as the model usual input.
  • bn_folding (List): List of tuple (nn.Conv2d, nn.BatchNorm2d) to be fused. If None it tries to evaluate them given the model. Default None.
  • fuse_bn (bool): If True, fuse the conv-bn tuple.
  • pinned_out (List): List of nn.Modules which output needs to remain of the original shape (e.g. layers related to a residual connection with a sum operation).

Minimal working example

import torch
from torchvision import models
from simplify import simplify

model = models.resnet18()

# Apply some pruning strategy or load a pruned checkpoint

dummy_input = torch.zeros(1, 3, 224, 224)  # Tensor shape is that of a standard input for the given model
simplified_model = simplify(model, dummy_input)

Submodules

The simplify function is composed of three different submodules: fuse, propagate and remove. Each module can be used independently as needed.

fuse

Fuses adjacent Conv (or Linear) and BatchNorm layers.

propagate

Propagates non-zero bias of pruned neurons to remaining neurons of the next layers.

remove

Removes zeroed neurons from the architecture.


Citing

If you use this software for research or application purposes, please use the following citation:

@article{bragagnolo2021simplify,
  title = {Simplify: A Python library for optimizing pruned neural networks},
  journal = {SoftwareX},
  volume = {17},
  pages = {100907},
  year = {2022},
  issn = {2352-7110},
  doi = {https://doi.org/10.1016/j.softx.2021.100907},
  url = {https://www.sciencedirect.com/science/article/pii/S2352711021001576},
  author = {Andrea Bragagnolo and Carlo Alberto Barbano},
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

torch-simplify-1.1.4.tar.gz (12.9 kB view details)

Uploaded Source

Built Distribution

torch_simplify-1.1.4-py3-none-any.whl (16.5 kB view details)

Uploaded Python 3

File details

Details for the file torch-simplify-1.1.4.tar.gz.

File metadata

  • Download URL: torch-simplify-1.1.4.tar.gz
  • Upload date:
  • Size: 12.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.16

File hashes

Hashes for torch-simplify-1.1.4.tar.gz
Algorithm Hash digest
SHA256 945a2448b79dd90f8d9d671028a235fa73be71e573f993e9bfe454ddf307a7cc
MD5 00065fb659d25a19ee4b51b243242cef
BLAKE2b-256 a54feb5e00a46092152bd7b68cf210aa714108ffa3d563b8145e14967da3894b

See more details on using hashes here.

File details

Details for the file torch_simplify-1.1.4-py3-none-any.whl.

File metadata

File hashes

Hashes for torch_simplify-1.1.4-py3-none-any.whl
Algorithm Hash digest
SHA256 f7dc4efccacba0b0a67ba791424f49a9e09269cf54216b1a27b0767b79968135
MD5 cd0d10cceac66e762966dba79c91c09d
BLAKE2b-256 352dec9d2cdea80d60c0dfda6a5c64abf924f3cb75625f7d02dcf6629f66d744

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page