Multi-perturbation Shapley value Analysis (MSA)
Project description
TLDR: A Game theoretical approach for calculating the contribution of each element of a system (here network models of the brain) to a system-wide description of the system. The classic neuroscience example: How much each brain region is causally relevant to an arbitrary cognitive function.
Motivation & such:
MSA is developed by Keinan and colleagues back in 2004, the motivation for them was to have a causal picture of the system by lesioning its elements. The method itself is not new, if not the first, it was among one of the earliest ones used by neuroscientists to understand the brain. The reasoning is quite simple, let us study broken systems to see what's missing both from the brain and the behavior (or cognition) and assume that region was causally necessary for the emergence of that cognitive/behavioral state. What MSA does is to see this necessity as contribution. If the brain region is indeed the seed of this cognitive function (whatever this means) then its contribution should be very high while other regions will have near zero contribution. Having this in mind then we can see the whole scenario as a cooperative game in which a coalition of players work together and obtain some divisible outcome, then the question is quite the same. How to divide the outcome to the players in a "fair" way such that the most "important" player gets the biggest chunk. Shapley value is then that chunk! It is the result of a mathematically rigorous and axiomatic procedure that derives who should get how much from all possible combinations of coalitions and all ordering in which players can enter the game. Translating it to neuroscience, it derives a ranking of contributions from a dataset of all possible combinations of lesions. This means 2N lesions (assuming lesions are binary, either perturbed or not), which N is the number of brain regions.
As you probably noticed this won't be feasible to calclulate as for example, it means a total number of 4,503,599,627,370,496 lesion combinations, assuming the brain is organized as Broadmann said, i.e., with 52 regions. So we estimate! For a more detailed description visit:
And our own recent work Fakhar K, Hilgetag CC. Systematic perturbation of an artificial neural network: A step towards quantifying causal contributions in the brain. PLoS Comput Biol. 2022;18: e1010250. doi:10.1371/journal.pcbi.1010250
Installation:
The easiest way is to pip install msapy
, This package is tested on Python 3.9
to Python 3.11
. Other versions might not work.
How it works:
Here you can see a schematic representation of how the algorithm works (interested in math instead? check the papers above). Briefly, all MSA needs from you is a list of players and a game function. The players can be your nodes, for example, brain regions or indices in a connectivity matrix, or links between them as tuples. It then shuffles them to produce N orderings in which they can join the game. This can end with repeating permutations if the set is small but that's fine don't worry! MSA then produces a "combination space" in which it produces all the combinations the player should form coalitions. Then it uses your game function and fills the contributions of those coalitions. The last step is to perform a Shapley integration and isolate each player's contribution in that given permutation. Repeating this for all permutations produces a contribution table (shapley table) and you'll get your shapley values by averaging over permutations so the end result is a value per element/player. To get a better grasp of how this works in code, check the minimal example in the examples folder.
How it works in Python:
I tried to make the package compact and easy-to-use but still there are a few things to keep in mind. Please take a look at the examples but just to give a flavor let's start working with the set ABCD as we have in the above picture.
Importing will be just:
from msapy import msa, utils as ut
Then we define some elements and generate the permutation space:
nodes = ['A', 'B', 'C', 'D']
permutation_space = msa.make_permutation_space(n_permutations=1000, elements=nodes)
This results in a list of tuples, our permutation space that has 1000 permutations in it, here are the top 5 ones:
[('D', 'C', 'A', 'B'),
('A', 'D', 'C', 'B'),
('D', 'A', 'B', 'C'),
('D', 'B', 'C', 'A'),
('A', 'D', 'C', 'B')]
Then we use this to produce our combination space:
combination_space = msa.make_combination_space(permutation_space=permutation_space)
And a quick look of what's inside:
[frozenset({'D'}),
frozenset(),
frozenset({'C', 'D'}),
frozenset({'A', 'C', 'D'}),
frozenset({'A', 'B', 'C', 'D'}),
frozenset({'A'}),
frozenset({'A', 'D'}),
frozenset({'A', 'B', 'D'}),
frozenset({'B', 'D'}),
frozenset({'B', 'C', 'D'}),
frozenset({'C'}),
frozenset({'A', 'B'}),
frozenset({'A', 'B', 'C'}),
frozenset({'A', 'C'}),
frozenset({'B', 'C'}),
frozenset({'B'})]
As you can see eventhough the permutation space has 1000 permutations, the combination space is exhausted because the total number of possible combinations is 24 or 16. Now here's the trick, we need to assign values to these combinations (coalitions) by keeping them intact while every other element is perturbed. In other words, the contribution of coalition {'B', 'C'}
is isolated if we lesion {'A', 'D'}
before playing the game. So what we do (and is not in the figure above) is to produce the "complement space" of the combination space:
complement_space = msa.make_complement_space(combination_space=combination_space, elements=nodes)
that is the difference of what's in the combination space in that coalition and what is not:
[('C', 'B', 'A'),
('C', 'D', 'B', 'A'),
('B', 'A'),
('B',),
(),
('C', 'D', 'B'),
('C', 'B'),
('C',),
('C', 'A'),
('A',),
('D', 'B', 'A'),
('C', 'D'),
('D',),
('D', 'B'),
('D', 'A'),
('C', 'D', 'A')]
As you can see, for example when combination is {'D'}
the corresponding complement is ('C', 'B', 'A')
. Note the difference in types, combination space is an OrderedSet
of frozenset
s so the Shapley value calculations are quicker while complement space is an OrderedSet
of Tuples
So handling it in your objective function is easier. Speaking of, let's make the worst objective function that just produces random values regardless of what's what (see the example on ground-truth models.ipynb
for a more elaborate version.)(see the example on ground-truth models.ipynb
for a more elaborate version.)
def rnd(complements):
return np.random.randint(1, 10)
We'll next play the games and aquire the contributions as follows:
contributions, lesion_effects = msa.take_contributions(elements=nodes,
combination_space=combination_space,
complement_space=complement_space,
objective_function=rnd)
Both contributions
and lesion_effects
are the same values just addressed differently. For example, if the contribution of coalition {'B', 'C'}
is 5 points then you can also say the effect of lesioning coalition {'A', 'D'}
is 5 points. This by itself is not that informative but if you know the contribution of the grand coalition (intact system) then you can claim that the effect of lesioning {'A', 'D'}
is a drop of some performance from x to 5.
Lastly, you can calculate Shapley values like:
import msa
shapley_table = msa.get_shapley_table(contributions=contributions, permutation_space=permutation_space)
Which gives you a ShapleyTable
data structure which is a wrapper around pd.DataFrame
to work with.
The Interface:
To make things easier, msa comes with an interface function:
shapley_table, contributions, lesion_effects = msa.interface(multiprocessing_method='joblib',
elements=regions,
n_permutations=1000,
objective_function=rnd,
n_parallel_games=-1,
random_seed=1)
For this one, all you have to do is to provide your elements, the objective function, and specify some parameters. For example, you can choose between two different multiprocessing toolboxes joblib
and ray
to distribute msa.take_contributions
over n_parallel_games
. Specifying a random_seed
is encouraged for reproducibility but the default is None
.
TODO (Interested in Contributing?):
- More estimation methods, for example see: amiratag/neuronshapley.
- GPU and HPC compatibilty
- Providing built-in objective functions for common use-cases.
- Improved documentation
- More Tests
Cite:
@misc{MSA,
author = {Kayson Fakhar and Shrey Dixit},
title = {MSA: A compact Python package for Multiperturbation Shapley value Analysis.},
year = {2021},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/kuffmode/msa}},
}
Acknowledgement:
I thank my good friend and Python mentor Fabrizio Damicelli whom I learned a lot from. Without him this package would be a disaster to look at.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file msapy-1.7.2.tar.gz
.
File metadata
- Download URL: msapy-1.7.2.tar.gz
- Upload date:
- Size: 25.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.1.0 CPython/3.12.4
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 7506e537a2f6acfb6698cecda73c005ae0503b96759bac2cc70cbc7a28df7f1c |
|
MD5 | 30290239e4453e23e40ee2fcfd9734ce |
|
BLAKE2b-256 | a2a881b8cc2d3303b8c55f8e2b1b7169942e81f4bcf703139de19754e5fe8659 |
File details
Details for the file msapy-1.7.2-py3-none-any.whl
.
File metadata
- Download URL: msapy-1.7.2-py3-none-any.whl
- Upload date:
- Size: 24.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.1.0 CPython/3.12.4
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 4db258dccd6499b75180bb4b70ad49bc091be842bf2f4e060257f04d017de963 |
|
MD5 | b0669178b5e2701704503a2cbb2d2a7f |
|
BLAKE2b-256 | 994a1a7c82a5adbd4081e3023891718cfd8d6da6bc40128d5c9425ec8065fe2e |