Python toolbox to create adversarial examples that fool neural networks
Foolbox is a Python toolbox to create adversarial examples that fool neural networks. It requires
pip install foolbox
We test using Python 2.7, 3.5 and 3.6. Other Python versions might work as well. We recommend using Python 3!
Documentation is available on readthedocs: http://foolbox.readthedocs.io/
import foolbox import keras from keras.applications.resnet50 import ResNet50, preprocess_input # instantiate model keras.backend.set_learning_phase(0) kmodel = ResNet50(weights='imagenet') fmodel = foolbox.models.KerasModel(kmodel, bounds=(0, 255), preprocess_fn=preprocess_input) # get source image and label image, label = foolbox.utils.imagenet_example() # apply attack on source image attack = foolbox.attacks.FGSM(fmodel) adv_img = attack(image=image, label=label)
Interfaces for a range of other deeplearning packages such as TensorFlow, PyTorch and Lasagne are available, e.g.
model = foolbox.models.TensorFlowModel(images, logits, bounds=(0, 255)) model = foolbox.models.PyTorchModel(torchmodel, bounds=(0, 255), num_classes=1000) # etc.
Different adversarial criteria such as Top-k, specific target classes or target probability levels can be passed to the attack, e.g.
criterion = foolbox.criteria.TargetClass(22) attack = foolbox.attacks.FGSM(fmodel, criterion)
Feature requests and bug reports
We welcome feature requests and bug reports. Just create a new issue on GitHub.
Depending on the nature of your question feel free to post it as an issue on GitHub, or post it as a question on Stack Overflow using the
foolbox tag. We will try to monitor that tag but if you don’t get an answer don’t hesitate to contact us.
Foolbox is a work in progress and any input is welcome.
If you find Foolbox useful for your scientific work, please consider citing it in resulting publications. We will soon publish a technical paper and will provide the citation here.