Skip to main content

Tile-based inference for segmentation of large images.

Project description

Conflux Segmentation

A Python library for tile-based inference for segmentation of large images.

Assuming you have a segmentation model that operates on tiles (e.g. 512 x 512), this library provides the plumbing to apply that model on a large image -- handling the padding, striding, and blending required.

Installation

pip install conflux-segmentation

Usage

The main Segmenter class assumes that the underlying tile-based segmenter outputs a multidimensional array of shape N x K x H x W where H and W are the height and width of a tile (e.g. 512), N is the batch size, and K is the output dimension (e.g. 1 for binary and > 1 for multiclass or multilabel).

Below we show an example of binary segmentation, although multiclass and multilabel are also supported. In this case, we assume the tile model outputs logits, so we specify "sigmoid" for the activation.

First, construct the Segmenter:

For PyTorch (e.g. with Segmentation Models PyTorch):

# $ pip install segmentation-models-pytorch
import segmentation_models_pytorch as smp
import torch
from conflux_segmentation import Segmenter

net = smp.Unet(encoder_name="tu-mobilenetv3_small_100", encoder_weights=None, activation=None)
net.load_state_dict(torch.load("/path/to/weights", weights_only=True))
net.eval()
segmenter = Segmenter.from_torch(net, activation="sigmoid")
# Alternatively, if your model already has a Sigmoid layer at the end:
# import torch.nn as nn
# sigmoid_net = nn.Sequential(net, nn.Sigmoid()).eval()
# segmenter = Segmenter.from_torch(net)

Or, for ONNX Runtime:

import onnxruntime as ort
from conflux_segmentation import Segmenter

session = ort.InferenceSession("/path/to/model.onnx")
segmenter = Segmenter.from_onnx(session, activation="sigmoid")

Then, to segment a large image:

# $ pip install opencv-python-headless
import cv2

# H x W x 3 image array of np.uint8
image = cv2.cvtColor(cv2.imread("/path/to/large/image"), cv2.COLOR_BGR2RGB)

result = segmenter(image).to_binary()
# H x W boolean array
mask = result.get_mask()
assert mask.shape == image.shape[:2]
assert (mask == True).sum() + (mask == False).sum() == mask.size

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

conflux_segmentation-0.4.2.tar.gz (9.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

conflux_segmentation-0.4.2-py3-none-any.whl (13.6 kB view details)

Uploaded Python 3

File details

Details for the file conflux_segmentation-0.4.2.tar.gz.

File metadata

  • Download URL: conflux_segmentation-0.4.2.tar.gz
  • Upload date:
  • Size: 9.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.2.1 CPython/3.12.3 Linux/6.11.0-1018-azure

File hashes

Hashes for conflux_segmentation-0.4.2.tar.gz
Algorithm Hash digest
SHA256 d22074fe98539657bd8a42013932f48312d10d7e8ea663ea47ba10f7f7a9f8fc
MD5 3ef046f4d993a65aa91a408ec7d6cba4
BLAKE2b-256 7435c9d30b79f40cde8b3589ac56104d494bed5d418ae18f340b250452831b9b

See more details on using hashes here.

File details

Details for the file conflux_segmentation-0.4.2-py3-none-any.whl.

File metadata

  • Download URL: conflux_segmentation-0.4.2-py3-none-any.whl
  • Upload date:
  • Size: 13.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.2.1 CPython/3.12.3 Linux/6.11.0-1018-azure

File hashes

Hashes for conflux_segmentation-0.4.2-py3-none-any.whl
Algorithm Hash digest
SHA256 d62278a3ae311f192e80e7608c331a26a6aeaf4730f973ef9b1b500e913e3ee5
MD5 840ce7d2dde8d09f6edefb897fe8b7f6
BLAKE2b-256 967e1892235fc8adb19c06a39920cc791dc710e46879413b6759389ef10f0af9

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page