Skip to main content

Bound propagation in Pytorch

Project description

Bound propagation

Linear and interval bound propagation in Pytorch with easy-to-use API and GPU support. Initially made as an alternative to the original CROWN implementation which featured only Numpy, lots of for-loops, and a cumbersome API. This library is comparable to auto_LiRPA but differs in design philosophy - auto_LiRPA parses the ONNX computation graph, but not all neural networks are representable in ONNX. For this library, we instead assume that the network is structured in a tree of nn.Modules, which we map to BoundModules using a visitor and abstract factory pattern for extensibility (see New Modules below).

To install:

pip install bound-propagation

Supported bound propagation methods:

For the examples below assume the following network definition:

from torch import nn
from bound_propagation import BoundModelFactory, HyperRectangle

class Network(nn.Sequential):
    def __init__(self, *args):
        if args:
            # To support __get_index__ of nn.Sequential when slice indexing
            # CROWN (and implicitly CROWN-IBP) is doing this underlying
            super().__init__(*args)
        else:
            in_size = 30
            classes = 10

            super().__init__(
                nn.Linear(in_size, 16),
                nn.Tanh(),
                nn.Linear(16, 16),
                nn.Tanh(),
                nn.Linear(16, classes)
            )

net = Network()

factory = BoundModelFactory()
net = factory.build(net)

The method also works with nn.Sigmoid and nn.ReLU, and the three custom layers Residual, Cat, and Parallel.

Interval bounds

To get interval bounds for either IBP, CROWN, or CROWN-IBP:

x = torch.rand(100, 30)
epsilon = 0.1
input_bounds = HyperRectangle.from_eps(x, epsilon)

ibp_bounds = net.ibp(input_bounds)
crown_bounds = net.crown(input_bounds).concretize()
crown_ibp_bounds = net.crown_ibp(input_bounds).concretize()

Linear bounds

To get linear bounds for either CROWN or CROWN-IBP:

x = torch.rand(100, 30)
epsilon = 0.1
input_bounds = HyperRectangle.from_eps(x, epsilon)

crown_bounds = net.crown(input_bounds)
crown_ibp_bounds = net.crown_ibp(input_bounds)

If lower or upper bounds are not needed, then you can add bound_lower=False or bound_lower=True to avoid the unnecessary computation. This also works for interval bounds with CROWN and CROWN-IBP. IBP accepts the two parameters for compatibility but ignores them since both bounds are necessary for the forward propagation.

New Modules

To show how to design a new module, we use a residual module as a running example (see residual.py for the full code). Adding a new module revolves around the assumption that each module as one input and one output, but a module can have submodules such as the case for residual.

class Residual(nn.Module):
    def __init__(self, subnetwork):
        super().__init__()
        self.subnetwork = subnetwork

    def forward(self, x):
        return self.subnetwork(x) + x

Next, we build the corresponding BoundModule. The first parameter of the BoundModule is the nn.Module, in this case Residual, and the second is a BoundModelFactory. Why factory? Because we do not know the structure of the subnetwork, and this could contain new network types defined by users of this library. So the factory allows us to defer defining how to construct the BoundModule for the subnetwork, and the construction happens by calling factory.build(subnetwork) (possibly multiple times if there are multiple subnetworks).

class BoundResidual(BoundModule):
    def __init__(self, model, factory, **kwargs):
        super().__init__(model, factory)
        self.subnetwork = factory.build(model.subnetwork)

    def propagate_size(self, in_size):
        out_size = self.subnetwork.propagate_size(in_size)
        assert in_size == out_size

        return out_size

A small note here is the need to implement propagate_size - some modules may need to know the input and output size without this being available on the nn.Module. Hence we propagate the size, and even though we know the output size from the input size for BoundResidual, we still propagate through the subnetwork to store this info throughout the bound graph.

Next, we consider linear relaxation for linear bounds - if you only use IBP, you can skip these. BoundModules must implement need_relaxation as a property and clear_relaxation as a method, and if the module requires relaxation (i.e. not linear and no submodules) it must implement backward_relaxation too. Note here that we let subnetwork handle the need_relaxation since Residual itself does not need it but might contain some activation function. Similarly, we refer to subnetwork for clear_relaxation. The backward_relaxation method is for adding relaxations to the bound graph using CROWN where the top-level function continually calls backward_relaxation until no more relaxation is needed as determined by need_relaxation. In this case the method may be called more than once since the subnetwork may have multiple layers needing relaxations, e.g. nn.Sequential with multiple layers of activation functions. If a module needs relaxation, it starts the backward relaxation chain with identity linear bounds (and zero bias accumulation) and self as the return tuple. The return signature of backward_relaxation is linear_bounds, relaxation_module since the top-level method needs a way of referring to the module being relaxed.

    @property
    def need_relaxation(self):
        return self.subnetwork.need_relaxation

    def clear_relaxation(self):
        self.subnetwork.clear_relaxation()

    def backward_relaxation(self, region):
        assert self.subnetwork.need_relaxation
        return self.subnetwork.backward_relaxation(region)

Finally, we get to the IBP forward and CROWN backward propagations themselves. IBP forward is straightforward using interval arithmetic with the exception that you must call bounds.region as the first parameter on IntervalBounds (and LinearBounds too). If not the memory usage will be excessive. IBP forward must also propagate save_relaxation and if a relaxation is needed for a module the relaxation must be constructed when IBP forward is called - this is to support CROWN-IBP. crown_backward has two components: updating linear bounds backwards and potentially adding to the bias accumulator. The structure depends largely on how the nn.Module is, but addition can be done as shown below for Residual. I recommend that you look at activation.py, linear.py, and cat.py to see how design crown_backward. Note that lower and upper will never interact, and that both can be None because we can avoid unnecessary computations if either is not needed.

    def ibp_forward(self, bounds, save_relaxation=False):
        residual_bounds = self.subnetwork.ibp_forward(bounds, save_relaxation=save_relaxation)

        return IntervalBounds(bounds.region, bounds.lower + residual_bounds.lower, bounds.upper + residual_bounds.upper)
    
    def crown_backward(self, linear_bounds):
        residual_linear_bounds = self.subnetwork.crown_backward(linear_bounds)

        if linear_bounds.lower is None:
            lower = None
        else:
            lower = (linear_bounds.lower[0] + residual_linear_bounds.lower[0], residual_linear_bounds.lower[1])

        if linear_bounds.upper is None:
            upper = None
        else:
            upper = (linear_bounds.upper[0] + residual_linear_bounds.upper[0], residual_linear_bounds.upper[1])

        return LinearBounds(linear_bounds.region, lower, upper)

The last task is to make a factory and register your new module (alternatively subclass the factory and add to the dictionary of nn.Modules and BoundModules).

factory = BoundModelFactory()
factory.register(Residual, BoundResidual)

Authors

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

bound_propagation-0.2.13.tar.gz (28.3 kB view hashes)

Uploaded Source

Built Distribution

bound_propagation-0.2.13-py3-none-any.whl (30.2 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page