Skip to main content

A Toolbox for the Evaluation of Explanations

Project description

Our AI generated logo. Comes from the prompt: 'logo of a t, inspired by an AI that is fair and responsible.'

teex: a toolbox for evaluating XAI explanations

PyPI Version Open GitHub Issues codecov Build Status Documentation Status

A Python toolbox for the evaluation of machine learning explanations.

This project aims to provide a simple way of evaluating individual black box explanations against ground truth. Moreover, it contains a collection of easy-to-access datasets with available g.t. explanations.

Installation

The teex package is on PyPI. To install it, simply run

pip install teex

teex is compatible with Python 3.8 and 3.9.

Documentation

teex's documentation and API reference can be found on Read The Docs.

Usage overview

teex is divided into subpackages, one for each explanation type. Each subpackage contains two modules, focused on two distinct functionalities:

  • eval: contains evaluation methods for that particular explanation type. For every subpackage, there is one high-level function to easily compute all the available metrics for an arbitrary number of explanations.
  • data: contains data classes with available g.t. explanations of that particular explanation type, both synthetic and real. All of them are objects that need to be instanced and, when sliced, will return the data, the target and the ground truth explanations, respectively.

Evaluation (with feature importance as an example)

What are feature importance vectors? They are vectors with one entry per feature. Each entry contains a weight that represents a feature's importance for the observation's outcome. Weights are usually in the range [-1, 1].

Suppose that we have a dataset with available g.t. explanations (gtExps) and a model trained with it (model):

from teex.featureImportance.eval import feature_importance_scores

# get individual feature importance explanations with any method
predictedExps = get_explanations(model, X)

# evaluate predicted explanations against ground truths
feature_importance_scores(gtExps, predictedExps, metrics=['fscore', 'cs', 'auc'])

This basic syntax is followed by the main evaluation APIs of all 4 explanation types:

  • Feature Importance: feature_importance_scores
  • Saliency Maps: saliency_map_scores
  • Decision Rules: rule_scores
  • Word Importance: word_importance_scores

Other functionalities are included in each evaluation module. More about each explanation type can be found in the example notebooks and the documentation.

Metrics supported:

Metrics available as of v1.0.0 are

  • Feature Importance
    • Cosine Similarity: similarity between the two vectors is measured in an inner product space in terms of orientation.
    • ROC AUC: where the ground truth is binarized in order for it to represent a class and the predicted vector entries are interpreted as classification scores or likelihood.
    • F1 Score: where both ground truth and prediction are binarized according to a user-defined threshold.
    • Precision: g.t. and prediction treated as in F1 Score
    • Recall: g.t. and prediction treated as in F1 Score
  • Saliency Maps
    • Same metrics as in feature importance. Each pixel in an image is considered to be a feature.
  • Decision Rules
    • Complete Rule Quality: Proportion of lower and upper bounds in a rule explanation whose that are eps-close to the respective lower and upper bounds (same feature) in the ground truth rule explanation amongst those that are not infinity.
    • All metrics in feature importance, where a transformation of the rule into feature importance vectors is performed first. See doc. for details.
  • Word Importance:
    • All metrics in feature importance, where a vocabulary is considered the feature space and a word importance explanation may or may not contain words from the vocabulary.

Note how in teex, feature importance vectors are a universal representation: we 'translate' all other explanation types to feature importance vectors to allow a wider metric space.

Data

teex also provides an easy way to get and use data with available ground truth explanations. It contains real datasets and can generate synthetic ones. All of them are instanced as objects, and can be sliced as usual. For example:

from teex.saliencyMap.data import Kahikatea

X, y, exps = Kahikatea()[:]

downloads and assigns data from the Kahikatea dataset:

drawing drawing

Fig. 1 A Kahikatea dataset sample.

Other datasets, such as CUB-200-2011 and the Oxford-IIIT Pet Dataset, are available on teex, with over 19000 images and 230 distinct classes:

drawing drawing

Fig. 2 A CUB-200-2011 dataset sample.

drawing drawing

Fig. 3 An Oxford-IIIT Pet Dataset sample.

Synthetic datasets can also be easily generated:

from teex.saliencyMap.data import SenecaSM

X, y, exps = SenecaSM()[:]

drawing drawing

Fig. 4 Artificial image and its g.t. saliency map explanation.

Tutorials and demos



Saliency maps


Feature importance vectors


Decision rules


Word importance vectors

Contributing

There is still work to do and we would really appreciate your help. Before contributing to teex, please take a moment to read the manual.

Acknowledgements

This work has been made possible by the University of Waikato under the scope of the TAIAO project.

drawing drawing drawing

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

teex-1.1.1.tar.gz (50.3 kB view details)

Uploaded Source

Built Distribution

teex-1.1.1-py3-none-any.whl (57.1 kB view details)

Uploaded Python 3

File details

Details for the file teex-1.1.1.tar.gz.

File metadata

  • Download URL: teex-1.1.1.tar.gz
  • Upload date:
  • Size: 50.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.9.6

File hashes

Hashes for teex-1.1.1.tar.gz
Algorithm Hash digest
SHA256 6bb6502fd044c59f8ab44bffc43c34297306e29cc82a0f2f7fd359634e467add
MD5 60d51ea16548f739855471f25f3a5649
BLAKE2b-256 60e73d25bf24f38b6a19881392f0cb98e2c5b31752f7c91d69e89bcd63026703

See more details on using hashes here.

File details

Details for the file teex-1.1.1-py3-none-any.whl.

File metadata

  • Download URL: teex-1.1.1-py3-none-any.whl
  • Upload date:
  • Size: 57.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.9.6

File hashes

Hashes for teex-1.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 b19d19c244316c399610bc0ecb41a37a3b735bd87342c8aa7bcec1de509c1895
MD5 85f7a5603b47bcf9da33c707fc1be2db
BLAKE2b-256 c2aa19c6f78a7e4a44a8bfcab006eab0238ee4c0db56ae21ed8e557378ae9843

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page