A metrics toolkit to evaluate neural network explanations.
Project description
A toolkit to evaluate neural network explanations
PyTorch and Tensorflow
Quantus is currently under active development so carefully note the Quantus release version to ensure reproducibility of your work.
News and Highlights! :rocket:
- Please see our latest release v0.2.0 which includes heavy changes to the API. Read more here!
- Offers more than 30+ metrics in 6 categories for XAI evaluation
- Supports different data types: image and time-series (NLP next up!)
- Flexible API: evaluate any PyTorch or Tensorflow model with your own customised explanation function(s)
- Different tutorials covering different datasets, models and explanation functions
- Latest metrics additions:
- Infidelity (Chih-Kuan, Yeh, et al., 2019)
- ROAD (Rong, Leemann, et al., 2022)
- Focus (Arias et al., 2022)
- Consistency (Dasgupta et al., 2022)
- Sufficiency (Dasgupta et al., 2022)
Citation
If you find this toolkit or its companion paper Quantus: An Explainable AI Toolkit for Responsible Evaluation of Neural Network Explanations interesting or useful in your research, use following Bibtex annotation to cite us:
@article{hedstrom2022quantus,
title={Quantus: An Explainable AI Toolkit for Responsible Evaluation of Neural Network Explanations},
author={Anna Hedström and
Leander Weber and
Dilyara Bareeva and
Franz Motzkus and
Wojciech Samek and
Sebastian Lapuschkin and
Marina M.-C. Höhne},
year={2022},
eprint={2202.06861},
archivePrefix={arXiv},
primaryClass={cs.LG}
}
When applying individual metrics of Quantus, please make sure to also properly cite the work of the original authors (as linked above).
Table of contents
Library overview
Simple visual comparison of eXplainable Artificial Intelligence (XAI) methods is often not sufficient to decide which explanation method works best as shown exemplary in Figure a) for four gradient-based methods — Saliency (Mørch et al., 1995; Baehrens et al., 2010), Integrated Gradients (Sundararajan et al., 2017), GradientShap (Lundberg and Lee, 2017) or FusionGrad (Bykov et al., 2021), yet it is a common practice for evaluation XAI methods in absence of ground truth data.
Therefore, we developed Quantus, an easy to-use yet comprehensive toolbox for quantitative evaluation of explanations — including 30+ different metrics. With Quantus, we can obtain richer insights on how the methods compare e.g., b) by holistic quantification on several evaluation criteria and c) by providing sensitivity analysis of how a single parameter e.g. the pixel replacement strategy of a faithfulness test influences the ranking of the XAI methods.
This project started with the goal of collecting existing evaluation metrics that have been introduced in the context of XAI research — to help automate the task of XAI quantification. Along the way of implementation, it became clear that XAI metrics most often belong to one out of six categories i.e., 1) faithfulness, 2) robustness, 3) localisation 4) complexity 5) randomisation or 6) axiomatic metrics (note, however, that the categories are oftentimes mentioned under different naming conventions e.g., 'robustness' is often replaced for 'stability' or 'sensitivity' and 'faithfulness' is commonly interchanged for 'fidelity'). The library contains implementations of the following evaluation metrics:
Faithfulness
quantifies to what extent explanations follow the predictive behaviour of the model (asserting that more important features play a larger role in model outcomes)- Faithfulness Correlation (Bhatt et al., 2020): iteratively replaces a random subset of given attributions with a baseline value and then measuring the correlation between the sum of this attribution subset and the difference in function output
- Faithfulness Estimate (Alvarez-Melis et al., 2018): computes the correlation between probability drops and attribution scores on various points
- Monotonicity Metric (Arya et al. 2019): starts from a reference baseline to then incrementally replace each feature in a sorted attribution vector, measuring the effect on model performance
- Monotonicity Metric (Nguyen et al, 2020): measures the spearman rank correlation between the absolute values of the attribution and the uncertainty in the probability estimation
- Pixel Flipping (Bach et al., 2015): captures the impact of perturbing pixels in descending order according to the attributed value on the classification score
- Region Perturbation (Samek et al., 2015): is an extension of Pixel-Flipping to flip an area rather than a single pixel
- Selectivity (Montavon et al., 2018): measures how quickly an evaluated prediction function starts to drop when removing features with the highest attributed values
- SensitivityN (Ancona et al., 2019): computes the correlation between the sum of the attributions and the variation in the target output while varying the fraction of the total number of features, averaged over several test samples
- IROF (Rieger at el., 2020): computes the area over the curve per class for sorted mean importances of feature segments (superpixels) as they are iteratively removed (and prediction scores are collected), averaged over several test samples
- Infidelity (Chih-Kuan, Yeh, et al., 2019): represents the expected mean square error between 1) a dot product of an attribution and input perturbation and 2) difference in model output after significant perturbation
- ROAD (Rong, Leemann, et al., 2022): measures the accuracy of the model on the test set in an iterative process of removing k most important pixels, at each step k most relevant pixels (MoRF order) are replaced with noisy linear imputations
- Sufficiency (Dasgupta et al., 2022): measures the extent to which similar explanations have the same prediction label
Robustness
measures to what extent explanations are stable when subject to slight perturbations of the input, assuming that model output approximately stayed the same- Local Lipschitz Estimate (Alvarez-Melis et al., 2018): tests the consistency in the explanation between adjacent examples
- Max-Sensitivity (Yeh et al., 2019): measures the maximum sensitivity of an explanation using a Monte Carlo sampling-based approximation
- Avg-Sensitivity (Yeh et al., 2019): measures the average sensitivity of an explanation using a Monte Carlo sampling-based approximation
- Continuity (Montavon et al., 2018): captures the strongest variation in explanation of an input and its perturbed version
- Consistency (Dasgupta et al., 2022): measures the probability that the inputs with the same explanation have the same prediction label
Localisation
tests if the explainable evidence is centered around a region of interest (RoI) which may be defined around an object by a bounding box, a segmentation mask or, a cell within a grid- Pointing Game (Zhang et al., 2018): checks whether attribution with the highest score is located within the targeted object
- Attribution Localization (Kohlbrenner et al., 2020): measures the ratio of positive attributions within the targeted object towards the total positive attributions
- Top-K Intersection (Theiner et al., 2021): computes the intersection between a ground truth mask and the binarized explanation at the top k feature locations
- Relevance Rank Accuracy (Arras et al., 2021): measures the ratio of highly attributed pixels within a ground-truth mask towards the size of the ground truth mask
- Relevance Mass Accuracy (Arras et al., 2021): measures the ratio of positively attributed attributions inside the ground-truth mask towards the overall positive attributions
- AUC (Fawcett et al., 2006): compares the ranking between attributions and a given ground-truth mask
- Focus (Arias et al., 2022): quantifies the precision of the explanation by creating mosaics of data instances from different classes
Complexity
captures to what extent explanations are concise i.e., that few features are used to explain a model prediction- Sparseness (Chalasani et al., 2020): uses the Gini Index for measuring, if only highly attributed features are truly predictive of the model output
- Complexity (Bhatt et al., 2020): computes the entropy of the fractional contribution of all features to the total magnitude of the attribution individually
- Effective Complexity (Nguyen at el., 2020): measures how many attributions in absolute values are exceeding a certain threshold
Randomisation
tests to what extent explanations deteriorate as inputs to the evaluation problem e.g., model parameters are increasingly randomised- Model Parameter Randomisation (Adebayo et. al., 2018): randomises the parameters of single model layers in a cascading or independent way and measures the distance of the respective explanation to the original explanation
- Random Logit Test (Sixt et al., 2020): computes for the distance between the original explanation and the explanation for a random other class
Axiomatic
assesses if explanations fulfill certain axiomatic properties- Completeness (Sundararajan et al., 2017): evaluates whether the sum of attributions is equal to the difference between the function values at the input x and baseline x'.
- Non-Sensitivity (Nguyen at el., 2020): measures whether the total attribution is proportional to the explainable evidence at the model output (and referred to as Summation to Delta (Shrikumar et al., 2017), Sensitivity-n (slight variation, Ancona et al., 2018) and Conservation (Montavon et al., 2018))
- Input Invariance (Kindermans et al., 2017): adds a shift to input, asking that attributions should not change in response (assuming the model does not)
Additional metrics will be included in future releases.
Disclaimers. It is worth noting that the implementations of the metrics in this library have not been verified by the original authors. Thus any metric implementation in this library may differ from the original authors. Further, bear in mind that evaluation metrics for XAI methods are often empirical interpretations (or translations) of qualities that some researcher(s) claimed were important for explanations to fulfill, so it may be a discrepancy between what the author claims to measure by the proposed metric and what is actually measured e.g., using entropy as an operationalisation of explanation complexity.
The first iteration has been developed primarily for image classification tasks, with attribution-based explanations in mind (which is a category of explanation methods that aim to assign an importance value to the model features and arguably, is the most studied kind of explanation). As a result, there will be both applications and explanation methods e.g., example-based methods where this library won't be applicable. Similarly, there is a couple of metrics that are popular but are considered out of scope for the first iteration of the library e.g., metrics that require re-training of the network e.g., RoAR (Hooker et al., 2018) and Label Randomisation Test (Adebayo et al., 2018) or rely on specifically designed datasets/ dataset modification e.g., Model Contrast Scores and Input Dependence Rate (Yang et al., 2019) and Attribution Percentage (Attr%) (Zhou et al., 2021).
Please read the user guidelines for further guidance on how to best use the library.
Installation
Quantus can be installed from PyPI
(this way assumes that you have either torch
or tensorflow
already installed on your machine).
pip install quantus
If you don't have torch
or tensorflow
installed, you can simply add the package you want and install it simultaneously.
pip install "quantus[torch]"
Or, alternatively for tensorflow
you run:
pip install "quantus[tensorflow]"
Additionally, if you want to use the basic explainability functionality such as quantus.explain
in your evaluations, you can run pip install "quantus[extras]"
(this step requires that either torch
or tensorflow
is installed).
To use Quantus with zennit
support, install in the following way: pip install "quantus[zennit]"
.
Alternatively, simply install requirements.txt (again, this requires that either torch
or tensorflow
is installed and won't include the explainability functionality to the installation):
pip install -r requirements.txt
Package requirements
python>=3.7.0
pytorch>=1.10.1
tensorflow==2.6.2
tqdm==4.62.3
Getting started
To use the library, you'll need a couple of ingredients; a model, some input data and labels (to be explained). In this example, we use torch
but we also support evaluation of tensorflow
models.
import quantus
import torch
import torchvision
# Enable GPU.
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
# Load a pre-trained LeNet classification model (architecture at quantus/helpers/models).
model = LeNet()
model.load_state_dict(torch.load("tutorials/assets/mnist"))
# Load datasets and make loaders.
test_set = torchvision.datasets.MNIST(root='./sample_data', download=True)
test_loader = torch.utils.data.DataLoader(test_set, batch_size=24)
# Load a batch of inputs and outputs to use for XAI evaluation.
x_batch, y_batch = iter(test_loader).next()
x_batch, y_batch = x_batch.cpu().numpy(), y_batch.cpu().numpy()
Next, we generate some explanations for some test set samples that we wish to evaluate using Quantus library.
import captum
from captum.attr import Saliency, IntegratedGradients
# Generate Integrated Gradients attributions of the first batch of the test set.
a_batch_saliency = Saliency(model).attribute(inputs=x_batch, target=y_batch, abs=True).sum(axis=1).cpu().numpy()
a_batch_intgrad = IntegratedGradients(model).attribute(inputs=x_batch, target=y_batch, baselines=torch.zeros_like(x_batch)).sum(axis=1).cpu().numpy()
# Save x_batch and y_batch as numpy arrays that will be used to call metric instances.
x_batch, y_batch = x_batch.cpu().numpy(), y_batch.cpu().numpy()
# Quick assert.
assert [isinstance(obj, np.ndarray) for obj in [x_batch, y_batch, a_batch_saliency, a_batch_intgrad]]
# You can use any function e.g., quantus.explain (not necessarily captum) to generate your explanations.
The qualitative aspects of the Saliency and Integrated Gradients explanations may look fairly uninterpretable - since we lack ground truth of what the explanations should be looking like, it is hard to draw conclusions about the explainable evidence that we see. So, to quantitatively evaluate the explanation we can apply Quantus. For this purpose, we may be interested in measuring how sensitive the explanations are to very slight perturbations. To this end, we can e.g., apply max-sensitivity by Yeh et al., 2019 to evaluate our explanations. With Quantus, we created two options for evaluation.
- Either evaluate the explanations in a one-liner - by calling the instance of the metric class.
# Define params for evaluation.
params_eval = {
"nr_samples": 10,
"perturb_radius": 0.1,
"norm_numerator": quantus.fro_norm,
"norm_denominator": quantus.fro_norm,
"perturb_func": quantus.uniform_noise,
"similarity_func": quantus.difference,
"img_size": 28,
"nr_channels": 1,
"normalise": False,
"abs": False,
"disable_warnings": True,
}
# Return max sensitivity scores in an one-liner - by calling the metric instance.
scores_saliency = quantus.MaxSensitivity(**params_eval)(model=model,
x_batch=x_batch,
y_batch=y_batch,
a_batch=a_batch_saliency,
**{"explain_func": quantus.explain,
"method": "Saliency",
"device": device})
- Or use
quantus.evaluate()
which is a high-level function that allow you to evaluate multiple XAI methods on several metrics at once.
import numpy as np
metrics = {"max-Sensitivity": quantus.MaxSensitivity(**params_eval),
}
xai_methods = {"Saliency": a_batch_saliency,
"IntegratedGradients": a_batch_intgrad}
results = quantus.evaluate(metrics=metrics,
xai_methods=xai_methods,
model=model,
x_batch=x_batch,
y_batch=y_batch,
agg_func=np.mean,
**{"explain_func": quantus.explain, "device": device})
# Summarise results in a dataframe.
df = pd.DataFrame(results)
df
When comparing the max-Sensitivity scores for the Saliency and Integrated Gradients explanations, we can conclude that in this experimental setting, Saliency can be considered less robust (scores 0.41 +-0.15std) compared to Integrated Gradients (scores 0.17 +-0.05std). To replicate this simple example please find a dedicated notebook: Getting started.
Tutorials
To get a more comprehensive view of the previous example, there is many types of analysis that can be done using Quantus. For example, we could use Quantus to verify to what extent the results - that Integrated Gradients "wins" over Saliency - are reproducible over different parameterisations of the metric e.g., by changing the amount of noise perturb_radius
or the number of samples to iterate over nr_samples
. With Quantus, we could further analyse if Integrated Gradients offers an improvement over Saliency also in other evaluation criteria such as faithfulness, randomisation and localisation.
For more use cases, please see notebooks in /tutorials
folder which includes examples such as
- Basic example all metrics: shows how to instantiate the different metrics for ImageNet
- Metrics' parameterisation sensitivity: explores how sensitive a metric could be to its hyperparameters
- Understand how explanations robustness develops during model training: looks into how robustness of gradient-based explanations change as model gets increasingly accurate in its predictions
- Compare explanation methods using qualitative-, quantitative- and sensitivity analysis: benchmarks explanation methods under different types of analysis: qualitative, quantitative and sensitivity
... and more.
Miscellaneous functionality
With Quantus, one can flexibly extend the library's functionality e.g., to adopt a customised explainer function explain_func
or to replace a function that perturbs the input perturb_func
with a user-defined one.
If you are replacing a function within the Quantus framework, make sure that your new function:
- returns the same datatype (e.g., np.ndarray or float) and,
- employs the same arguments (e.g., img=x, a=a) as the function you’re intending to replace.
Details on what datatypes and arguments that should be used for the different functions can be found in the respective function typing inquantus/helpers
. For example, if you want to replace similar_func
in your evaluation, you can do as follows.
import scipy
import numpy as np
def correlation_spearman(a: np.array, b: np.array, **kwargs) -> float:
"""Calculate Spearman rank of two images (or explanations)."""
return scipy.stats.spearmanr(a, b)[0]
def my_similar_func(a: np.array, b: np.array, **kwargs) -> float:
"""Calculate the similarity of a and b by subtraction."""
return a - b
# Simply initalise the metric with your own function.
metric = LocalLipschitzEstimate(similar_func=my_similar_func)
To evaluate multiple explanation methods over several metrics at once we user can leverage the evaluate
method in Quantus. There are also other miscellaneous functionality built-into Quantus that might be helpful:
# Interpret scores of a given metric.
metric_instance.interpret_scores
# Understand what hyperparameters of a metric to tune.
sensitivity_scorer.get_params
# To list available metrics (and their corresponding categories).
quantus.AVAILABLE_METRICS
# To list available explainable methods.
quantus.AVAILABLE_XAI_METHODS
# To list available perturbation functions.
quantus.AVAILABLE_SIMILARITY_FUNCTIONS
# To list available similarity functions.
quantus.AVAILABLE_PERTURBATION_FUNCTIONS
# To list available normalisation function.
quantus.AVAILABLE_NORMALISATION_FUNCTIONS
# To get the scores of the last evaluated batch.
metric_instance_called.last_results
# To get the scores of all the evaluated batches.
metric_instance_called.all_results
With each metric intialisation, warnings are printed to shell in order to make the user attentive to the hyperparameters of the metric which may have great influence on the evaluation outcome. If you are running evaluation iteratively you might want to disable warnings, then set:
disable_warnings = True
in the params of the metric initalisation. Additionally, if you want to track progress while evaluating your explanations set:
display_progressbar = True
If you want to return an aggreagate score for your test samples you can set the following hyperparameter:
return_aggregate = True
fow which you can specify an aggregate_func
e.g., np.mean
to use while aggregating the score for a given metric.
Contributing
If you would like to contribute to this project or add your metric to evaluate explanations please open an issue or submit a pull request.
Code Style
Code is written to follow PEP-8 and for docstrings we use numpydoc. We use flake8 for quick style checks and black for code formatting with a line-width of 88 characters per line.
Testing
Tests are written using pytest and executed together with codecov for coverage reports.
Workflow
Before creating a PR, double-check that the following tasks are completed:
- Run
black
to format source code e.g.,black quantus/helpers/INSERT_YOUR_FILE_NAME.py
- Run
flake8
for quick style checks e.g.,flake8 quantus/helpers/INSERT_YOUR_FILE_NAME.py
- Make
pytests
and add undertests/
folder (to install mandatory packages for testing runpip install -r requirements_text.txt
) - If the
pytests
include a new category of@pytest.mark
then add that category with description topytest.ini
- Run
pytest tests -v --cov-report term --cov-report html:htmlcov --cov-report xml --cov=quantus
to inspect that code coverage is maintained (we aim at ~100% code coverage for Quantus)
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file quantus-0.2.0.tar.gz
.
File metadata
- Download URL: quantus-0.2.0.tar.gz
- Upload date:
- Size: 116.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.8.0 pkginfo/1.8.2 readme-renderer/32.0 requests/2.26.0 requests-toolbelt/0.9.1 urllib3/1.26.7 tqdm/4.62.3 importlib-metadata/4.8.1 keyring/23.4.0 rfc3986/1.5.0 colorama/0.4.4 CPython/3.9.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 9da1c9115f89527461dfe82f33cd3e339e9156d0fac6c3ec4257acac22b0e4a8 |
|
MD5 | ac9b2cf1282efcd8a2aeabfb0459ea1e |
|
BLAKE2b-256 | 5042fdf5e6d4a5dd3e0ec89133f9271435a889eefbdf3bc68984d747960e17e7 |
File details
Details for the file quantus-0.2.0-py3-none-any.whl
.
File metadata
- Download URL: quantus-0.2.0-py3-none-any.whl
- Upload date:
- Size: 226.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.8.0 pkginfo/1.8.2 readme-renderer/32.0 requests/2.26.0 requests-toolbelt/0.9.1 urllib3/1.26.7 tqdm/4.62.3 importlib-metadata/4.8.1 keyring/23.4.0 rfc3986/1.5.0 colorama/0.4.4 CPython/3.9.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | fb22e2acb91274c854137ee3e074599da93775e209c7e52ce8290974729cc724 |
|
MD5 | 28308f8b864a2b9b95ae6d995f9a37f3 |
|
BLAKE2b-256 | cc3ca2a9b766669be390ef2c99c073166f744cbc79830b5907940ff8a257b868 |