PyTorch extensions for fast R&D prototyping and Kaggle farming
Project description
Pytorch-toolbelt
A pytorch-toolbelt
is a Python library with a set of bells and whistles for PyTorch for fast R&D prototyping and Kaggle farming:
What's inside
- Easy model building using flexible encoder-decoder architecture.
- Modules: CoordConv, SCSE, Hypercolumn, Depthwise separable convolution and more.
- GPU-friendly test-time augmentation TTA for segmentation and classification
- GPU-friendly inference on huge (5000x5000) images
- Every-day common routines (fix/restore random seed, filesystem utils, metrics)
- Losses: BinaryFocalLoss, Focal, ReducedFocal, Lovasz, Jaccard and Dice losses, Wing Loss and more.
- Extras for Catalyst library (Visualization of batch predictions, additional metrics)
Showcase: Catalyst, Albumentations, Pytorch Toolbelt example: Semantic Segmentation @ CamVid
Why
Honest answer is "I needed a convenient way to re-use code for my Kaggle career". During 2018 I achieved a Kaggle Master badge and this been a long path. Very often I found myself re-using most of the old pipelines over and over again. At some point it crystallized into this repository.
This lib is not meant to replace catalyst / ignite / fast.ai. Instead it's designed to complement them.
Installation
pip install pytorch_toolbelt
Showcase
Encoder-decoder models construction
from pytorch_toolbelt.modules import encoders as E
from pytorch_toolbelt.modules import decoders as D
class FPNSegmentationModel(nn.Module):
def __init__(self, encoder:E.EncoderModule, num_classes, fpn_features=128):
self.encoder = encoder
self.decoder = D.FPNDecoder(encoder.output_filters, fpn_features=fpn_features)
self.fuse = D.FPNFuse()
input_channels = sum(self.decoder.output_filters)
self.logits = nn.Conv2d(input_channels, num_classes,kernel_size=1)
def forward(self, input):
features = self.encoder(input)
features = self.decoder(features)
features = self.fuse(features)
logits = self.logits(features)
return logits
def fpn_resnext50(num_classes):
encoder = E.SEResNeXt50Encoder()
return FPNSegmentationModel(encoder, num_classes)
def fpn_mobilenet(num_classes):
encoder = E.MobilenetV2Encoder()
return FPNSegmentationModel(encoder, num_classes)
Compose multiple losses
from pytorch_toolbelt import losses as L
loss = L.JointLoss(L.FocalLoss(), 1.0, L.LovaszLoss(), 0.5)
Test-time augmentation
from pytorch_toolbelt.inference import tta
# Truly functional TTA for image classification using horizontal flips:
logits = tta.fliplr_image2label(model, input)
# Truly functional TTA for image segmentation using D4 augmentation:
logits = tta.d4_image2mask(model, input)
# TTA using wrapper module:
tta_model = tta.TTAWrapper(model, tta.fivecrop_image2label, crop_size=512)
logits = tta_model(input)
Inference on huge images:
import numpy as np
import torch
import cv2
from pytorch_toolbelt.inference.tiles import ImageSlicer, CudaTileMerger
from pytorch_toolbelt.utils.torch_utils import tensor_from_rgb_image, to_numpy
image = cv2.imread('really_huge_image.jpg')
model = get_model(...)
# Cut large image into overlapping tiles
tiler = ImageSlicer(image.shape, tile_size=(512, 512), tile_step=(256, 256), weight='pyramid')
# HCW -> CHW. Optionally, do normalization here
tiles = [tensor_from_rgb_image(tile) for tile in tiler.split(image)]
# Allocate a CUDA buffer for holding entire mask
merger = CudaTileMerger(tiler.target_shape, 1, tiler.weight)
# Run predictions for tiles and accumulate them
for tiles_batch, coords_batch in DataLoader(list(zip(tiles, tiler.crops)), batch_size=8, pin_memory=True):
tiles_batch = tiles_batch.float().cuda()
pred_batch = model(tiles_batch)
merger.integrate_batch(pred_batch, coords_batch)
# Normalize accumulated mask and convert back to numpy
merged_mask = np.moveaxis(to_numpy(merger.merge()), 0, -1).astype(np.uint8)
merged_mask = tiler.crop_to_orignal_size(merged_mask)
Advanced examples
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
File details
Details for the file pytorch_toolbelt-0.1.2.tar.gz
.
File metadata
- Download URL: pytorch_toolbelt-0.1.2.tar.gz
- Upload date:
- Size: 59.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.22.0 setuptools/41.0.1 requests-toolbelt/0.9.1 tqdm/4.32.2 CPython/3.6.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 4e292a9867c25b5844cf12be318eab89a05bac72fa115c19b433e51fe35a7abd |
|
MD5 | f8f666c49161c4c8d8f714e5de4755b2 |
|
BLAKE2b-256 | 7337c5cde0e02fec372ed77b3547c385a32d900cc6f02fbd050e8b6668240145 |