Generating global explanations from local ones.
Project description
GLocalX - Global through Local Explainability
Explanations come in two forms: local, explaining a single model prediction, and global, explaining all model predictions. The Local to Global (L2G) problem consists in bridging these two family of explanations. Simply put, we generate global explanations by merging local ones.
The algorithm
Local and global explanations are provided in the form of decision rules:
age < 40, income > 50000, status = married, job = CEO ⇒ grant loan application
This rule describes the rationale followed given by an unexplainable model to grant the loan application to an individual younger than 40 years-old, with an income above 50000$, married and currently working as a CEO.
Setup
git clone https://github.com/msetzu/glocalx/
cd glocalx
Dependencies are listed in requirements.txt
, a virtual environment is advised:
mkvirtualenv glocalx # optional but reccomended
pip3 install -r requirements.txt
Running the code
Installing from pypi does not give direct access to quickstart data, first download the data/
folder from the original repository.
Python interface
from tensorflow.keras.models import load_model
import logzero
from glocalx.glocalx import GLocalX, shut_up_tensorflow
from glocalx.models import Rule
# Set log profile: INFO for normal logging, DEBUG for verbosity
logzero.loglevel(logzero.logging.INFO)
shut_up_tensorflow()
# Load black box: optional! Use black_box = None to use the dataset labels
black_box = load_model('data/dummy/dummy_model.h5')
# Load data and header
data = genfromtxt('data/dummy/dummy_dataset.csv', delimiter=',', names=True)
features_names = data.dtype.names
tr_set = data.view(float).reshape(data.shape + (-1,))
# Load local explanations
local_explanations = Rule.from_json('data/dummy/dummy_rules.json', names=features_names)
# Create a GLocalX instance for `black_box`
glocalx = GLocalX(oracle=black_box)
# Fit the model, use batch_size=128 for larger datasets
glocalx = glocalx.fit(local_explanations, tr_set, batch_size=2, name='black_box_explanations')
# Retrieve global explanations by fidelity
alpha = 0.5
global_explanations = glocalx.rules(alpha, tr_set)
# Retrieve global explanations by fidelity percentile
alpha = 95
global_explanations = glocalx.rules(alpha, tr_set, is_percentile=True)
# Retrieve exactly `alpha` global explanations, `alpha/2` per class
alpha = 10
global_explanations = glocalx.rules(alpha, tr_set)
Command line interface
You can invoke GLocalX
from the api interface in api.py
:
> python3 api.py --help
Usage: api.py [OPTIONS] RULES TR
Options:
-o, --oracle PATH
--names TEXT Features names.
-cbs, --callbacks FLOAT Callback step, either int or float. Defaults
to 0.1
-m, --name TEXT Name of the log files.
--generate TEXT Number of records to generate, if given.
Defaults to None.
-i, --intersect TEXT Whether to use coverage intersection
('coverage') or polyhedra intersection
('polyhedra'). Defaults to 'coverage'.
-f, --fidelity_weight FLOAT Fidelity weight. Defaults to 1.
-c, --complexity_weight FLOAT Complexity weight. Defaults to 1.
-a, --alpha FLOAT Pruning factor. Defaults to 0.5
-b, --batch INTEGER Batch size. Set to -1 for full batch.
Defaults to 128.
-u, --undersample FLOAT Undersample size, to use a percentage of the
rules. Defaults to 1.0 (No undersample).
--strict_join Use to use high concordance.
--strict_cut Use to use the strong cut.
--global_direction Use to use the global search direction.
-d, --debug INTEGER Debug level.
--help Show this message and exit.
A minimal run simply requires a set of local input rules, a training set and (optionally) a black box:
python3.8 apy.py data/dummy/dummy_rules.json data/dummy/dummy_dataset.csv --oracle data/dummy/dummy_model.h5 --name dummy --batch 2
If you are interested, the folder data/dummy/
contains a dummy example.
You can run it with
python3.8 apy.py my_rules.json training_set.csv --oracle my_black_box.h5 \
--name my_black_box_explanations
The remaining hyperparameters are optional:
--names $names
Comma-separated features names--callbacks $callbacks
Callback step, either int or float. Callbacks are invoked every--callbacks
iterations of the algorithm--name $name
Name of the log files and output. The script dumps here all the additional info as it executes--generate $size
Number of synthetic records to generate, if you don't wish to use the provided training set--intersect $strategy
Intersection strategy: eithercoverage
orpolyhedra
. Defaults tocoverage
--fidelity_weight $fidelity
Fidelity weight to reward accurate yet complex models. Defaults to 1.--complexity_weight $complexity
Complexity weight to reward simple yet less accurate models. Defaults to 1.--alpha $alpha
Pruning factor. Defaults to 0.5--batch $batch
Batch size. Set to -1 for full batch. Defaults to 128.--undersample $pct
Undersample size, to use a percentage of the input rules. Defaults to 1.0 (all rules)--strict_join
to use a more stringentjoin
--strict_cut
to use a more stringentcut
--global_direction
Use to evaluate merges on the whole validation set--debug
Debug level: the higher, the less messages shown
Validation
GLocalX
outputs a set of rules (list of models.Rule
) stored in a $names.rules.glocalx.alpha=$alpha.json
file.
The evaluators.validate
function provides a simple interface for validation. If you wish to extend it, you can directly extend either the evaluators.DummyEvaluator
or evaluators.MemEvaluator
class.
Run on your own dataset
GLocalX has a strict format on input data. It accepts tabular datasets and binary classification tasks. You can find a dummy example for each of these formats in /data/dummy/
.
Local rules
We provide an integration with Lore rules through the loaders
module.
To convert Lore rules to GLocalX rules, use the provided loaders.lore.lore_to_glocalx(json_file, info_file)
function. json_file
should be the path to a json with Lore's rules, and info_file
should be the path to a JSON dictionary holding the class values (key class_names
) and a list of the features' names (key feature_names
).
You can find dummy examples of the info file in data/loaders/adult_info.json.
More examples on other local rule extractors to follow.
Rules [/data/dummy/dummy_rules.json
]
Local rules are to be stored in a JSON
format:
[
{"22": [30.0, 91.9], "23": [-Infinity, 553.3], "label": 0},
...
]
Each rule in the list is a dictionary with an arbitrary (greater than 2) premises. The rule prediction ({0, 1}) is stored in the key label
. Premises on features are stored according to their ordering and bounds: in the above, "22": [-Infinity, 91.9]
indicates the premise "feature number 22 has value between 30.0 and 91.9".
Black boxes [/data/dummy/dummy_model.h5
]
Black boxes (if used) are to be stored in a hdf5
format if given through command line. If given programmatically instead, it suffices that they implement the Predictor
interface:
class Predictor:
@abstractmethod
def predict(self, x):
pass
when called to predict numpy.ndarray:x
the predictor shall return its predictions in a numpy.ndarray
of integers.
Training data[/data/dummy/dummy_dataset.csv
]
Training data is to be stored in a csv, comma-separated format with features names as header. The classification labels should have feature name y
.
Docs and reference
You can find the software documentation in the /html/
folder and a powerpoint presentation on GLocalX can be found here.
You can cite this work with
@article{SETZU2021103457,
title = {GLocalX - From Local to Global Explanations of Black Box AI Models},
journal = {Artificial Intelligence},
volume = {294},
pages = {103457},
year = {2021},
issn = {0004-3702},
doi = {https://doi.org/10.1016/j.artint.2021.103457},
url = {https://www.sciencedirect.com/science/article/pii/S0004370221000084},
author = {Mattia Setzu and Riccardo Guidotti and Anna Monreale and Franco Turini and Dino Pedreschi and Fosca Giannotti},
keywords = {Explainable AI, Global explanation, Local explanations, Interpretable models, Open the black box},
}
Useful functions & Overrides
Callbacks
The fit()
function provides a callbacks
parameter to add any callbacks you desire to be invoked every callbacks_step
iterations. The callback should implement the callbacks.Callback
interface. You can find the set of parameters available to the callback in glocalx.GLocalX.fit()
.
Serialization and deserialization
You can dump to disk and load GLocalX
instances and their output with the Rule
object and the serialization
module:
from models import Rule
import serialization
rules_only_json = 'input_rules.json'
run_file = 'my_run.glocalx.json'
# Load input rules
rules = Rule.from_json(rules_only_json)
# Load GLocalX output
glocalx_output = serialization.load_run(run_file)
# Load a GLocalX instance from a set of rules, regardless of whether they come from an actual run or not!
# From a GLocalX run...
glocalx = serialization.load_glocalx(run_file, is_glocalx_run=True)
# From a
glocalx = serialization.load_glocalx(rules_only_json, is_glocalx_run=False)
Extending the merge
function
To override the merge function, simply extend the glocalx.GLocalX
object and override the merge
function with the following signature:
merge(self, A:set, B:set, x:numpy.ndarray, y:numpy.ndarray, ids:numpy.ndarray)
where A
and B
are the sets of models.Rule
you are merging, x
is the training data, y
are the training labels and ids
are the batch ids. The ids are used by the MemEvaluator
to store pre-computed results.
Extending the distance
function
The distance
between explanations is computed by the evaluators.Evaluator
objects. To override it, override either the evaluators.DummyEvaluator
or evaluators.MemEvaluator
object with the following signature:
distance(self, A:set, B:set, x:numpy.ndarray, ids:numpy.ndarray) -> numpy.ndarray
where A
, B
are the two (sets of) explanation(s), x
is the training data and ids
are the ids for the current batch.
Extending the merge/acceptance
Whether a merge is accepted or rejected is decided by the glocalx.GLocalX.accept_merge()
function with signature:
accept_merge(union:set, merge:set, **kwargs) -> bool
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file glocalx-0.0.9.tar.gz
.
File metadata
- Download URL: glocalx-0.0.9.tar.gz
- Upload date:
- Size: 52.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.3
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | ddb7d4eaf5607dd1dacdefa9c7d87b134af91e6a2a496a84f19f3f5424a31b8d |
|
MD5 | cfc976aff4e3ac3c0f29e4f95fa57abf |
|
BLAKE2b-256 | 949a7603b80d01d54cbf355458cad1ab6f59c25b22ecec25f551438885103179 |
File details
Details for the file glocalx-0.0.9-py3-none-any.whl
.
File metadata
- Download URL: glocalx-0.0.9-py3-none-any.whl
- Upload date:
- Size: 49.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.3
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | e1b36fe573b2958058b19c995d55c4fc6243255286acc5b48a082eddd84c0f44 |
|
MD5 | bf7228644b6e7ca7f65a2ce12da3b284 |
|
BLAKE2b-256 | 15d792b3b223ae17a192a89e9d55bbcc33e457137bd9d437bce2f23027248533 |