Skip to main content

A unified approach to explain the output of any machine learning model.

Project description


PyPI Conda License Tests Binder Documentation Status Downloads PyPI pyversions

SHAP (SHapley Additive exPlanations) is a game theoretic approach to explain the output of any machine learning model. It connects optimal credit allocation with local explanations using the classic Shapley values from game theory and their related extensions (see papers for details and citations).

Install

SHAP can be installed from either PyPI or conda-forge:

pip install shap
or
conda install -c conda-forge shap

Tree ensemble example (XGBoost/LightGBM/CatBoost/scikit-learn/pyspark models)

While SHAP can explain the output of any machine learning model, we have developed a high-speed exact algorithm for tree ensemble methods (see our Nature MI paper). Fast C++ implementations are supported for XGBoost, LightGBM, CatBoost, scikit-learn and pyspark tree models:

import xgboost
import shap

# train an XGBoost model
X, y = shap.datasets.california()
model = xgboost.XGBRegressor().fit(X, y)

# explain the model's predictions using SHAP
# (same syntax works for LightGBM, CatBoost, scikit-learn, transformers, Spark, etc.)
explainer = shap.Explainer(model)
shap_values = explainer(X)

# visualize the first prediction's explanation
shap.plots.waterfall(shap_values[0])

The above explanation shows features each contributing to push the model output from the base value (the average model output over the training dataset we passed) to the model output. Features pushing the prediction higher are shown in red, those pushing the prediction lower are in blue. Another way to visualize the same explanation is to use a force plot (these are introduced in our Nature BME paper):

# visualize the first prediction's explanation with a force plot
shap.plots.force(shap_values[0])

If we take many force plot explanations such as the one shown above, rotate them 90 degrees, and then stack them horizontally, we can see explanations for an entire dataset (in the notebook this plot is interactive):

# visualize all the training set predictions
shap.plots.force(shap_values[:500])

To understand how a single feature effects the output of the model we can plot the SHAP value of that feature vs. the value of the feature for all the examples in a dataset. Since SHAP values represent a feature's responsibility for a change in the model output, the plot below represents the change in predicted house price as the latitude changes. Vertical dispersion at a single value of latitude represents interaction effects with other features. To help reveal these interactions we can color by another feature. If we pass the whole explanation tensor to the color argument the scatter plot will pick the best feature to color by. In this case it picks longitude.

# create a dependence scatter plot to show the effect of a single feature across the whole dataset
shap.plots.scatter(shap_values[:, "Latitude"], color=shap_values)

To get an overview of which features are most important for a model we can plot the SHAP values of every feature for every sample. The plot below sorts features by the sum of SHAP value magnitudes over all samples, and uses SHAP values to show the distribution of the impacts each feature has on the model output. The color represents the feature value (red high, blue low). This reveals for example that higher median incomes improves the predicted home price.

# summarize the effects of all the features
shap.plots.beeswarm(shap_values)

We can also just take the mean absolute value of the SHAP values for each feature to get a standard bar plot (produces stacked bars for multi-class outputs):

shap.plots.bar(shap_values)

Natural language example (transformers)

SHAP has specific support for natural language models like those in the Hugging Face transformers library. By adding coalitional rules to traditional Shapley values we can form games that explain large modern NLP model using very few function evaluations. Using this functionality is as simple as passing a supported transformers pipeline to SHAP:

import transformers
import shap

# load a transformers pipeline model
model = transformers.pipeline('sentiment-analysis', return_all_scores=True)

# explain the model on two sample inputs
explainer = shap.Explainer(model)
shap_values = explainer(["What a great movie! ...if you have no taste."])

# visualize the first prediction's explanation for the POSITIVE output class
shap.plots.text(shap_values[0, :, "POSITIVE"])

Deep learning example with DeepExplainer (TensorFlow/Keras models)

Deep SHAP is a high-speed approximation algorithm for SHAP values in deep learning models that builds on a connection with DeepLIFT described in the SHAP NIPS paper. The implementation here differs from the original DeepLIFT by using a distribution of background samples instead of a single reference value, and using Shapley equations to linearize components such as max, softmax, products, divisions, etc. Note that some of these enhancements have also been since integrated into DeepLIFT. TensorFlow models and Keras models using the TensorFlow backend are supported (there is also preliminary support for PyTorch):

# ...include code from https://github.com/keras-team/keras/blob/master/examples/demo_mnist_convnet.py

import shap
import numpy as np

# select a set of background examples to take an expectation over
background = x_train[np.random.choice(x_train.shape[0], 100, replace=False)]

# explain predictions of the model on four images
e = shap.DeepExplainer(model, background)
# ...or pass tensors directly
# e = shap.DeepExplainer((model.layers[0].input, model.layers[-1].output), background)
shap_values = e.shap_values(x_test[1:5])

# plot the feature attributions
shap.image_plot(shap_values, -x_test[1:5])

The plot above explains ten outputs (digits 0-9) for four different images. Red pixels increase the model's output while blue pixels decrease the output. The input images are shown on the left, and as nearly transparent grayscale backings behind each of the explanations. The sum of the SHAP values equals the difference between the expected model output (averaged over the background dataset) and the current model output. Note that for the 'zero' image the blank middle is important, while for the 'four' image the lack of a connection on top makes it a four instead of a nine.

Deep learning example with GradientExplainer (TensorFlow/Keras/PyTorch models)

Expected gradients combines ideas from Integrated Gradients, SHAP, and SmoothGrad into a single expected value equation. This allows an entire dataset to be used as the background distribution (as opposed to a single reference value) and allows local smoothing. If we approximate the model with a linear function between each background data sample and the current input to be explained, and we assume the input features are independent then expected gradients will compute approximate SHAP values. In the example below we have explained how the 7th intermediate layer of the VGG16 ImageNet model impacts the output probabilities.

from keras.applications.vgg16 import VGG16
from keras.applications.vgg16 import preprocess_input
import keras.backend as K
import numpy as np
import json
import shap

# load pre-trained model and choose two images to explain
model = VGG16(weights='imagenet', include_top=True)
X,y = shap.datasets.imagenet50()
to_explain = X[[39,41]]

# load the ImageNet class names
url = "https://s3.amazonaws.com/deep-learning-models/image-models/imagenet_class_index.json"
fname = shap.datasets.cache(url)
with open(fname) as f:
    class_names = json.load(f)

# explain how the input to the 7th layer of the model explains the top two classes
def map2layer(x, layer):
    feed_dict = dict(zip([model.layers[0].input], [preprocess_input(x.copy())]))
    return K.get_session().run(model.layers[layer].input, feed_dict)
e = shap.GradientExplainer(
    (model.layers[7].input, model.layers[-1].output),
    map2layer(X, 7),
    local_smoothing=0 # std dev of smoothing noise
)
shap_values,indexes = e.shap_values(map2layer(to_explain, 7), ranked_outputs=2)

# get the names for the classes
index_names = np.vectorize(lambda x: class_names[str(x)][1])(indexes)

# plot the explanations
shap.image_plot(shap_values, to_explain, index_names)

Predictions for two input images are explained in the plot above. Red pixels represent positive SHAP values that increase the probability of the class, while blue pixels represent negative SHAP values the reduce the probability of the class. By using ranked_outputs=2 we explain only the two most likely classes for each input (this spares us from explaining all 1,000 classes).

Model agnostic example with KernelExplainer (explains any function)

Kernel SHAP uses a specially-weighted local linear regression to estimate SHAP values for any model. Below is a simple example for explaining a multi-class SVM on the classic iris dataset.

import sklearn
import shap
from sklearn.model_selection import train_test_split

# print the JS visualization code to the notebook
shap.initjs()

# train a SVM classifier
X_train,X_test,Y_train,Y_test = train_test_split(*shap.datasets.iris(), test_size=0.2, random_state=0)
svm = sklearn.svm.SVC(kernel='rbf', probability=True)
svm.fit(X_train, Y_train)

# use Kernel SHAP to explain test set predictions
explainer = shap.KernelExplainer(svm.predict_proba, X_train, link="logit")
shap_values = explainer.shap_values(X_test, nsamples=100)

# plot the SHAP values for the Setosa output of the first instance
shap.force_plot(explainer.expected_value[0], shap_values[0][0,:], X_test.iloc[0,:], link="logit")

The above explanation shows four features each contributing to push the model output from the base value (the average model output over the training dataset we passed) towards zero. If there were any features pushing the class label higher they would be shown in red.

If we take many explanations such as the one shown above, rotate them 90 degrees, and then stack them horizontally, we can see explanations for an entire dataset. This is exactly what we do below for all the examples in the iris test set:

# plot the SHAP values for the Setosa output of all instances
shap.force_plot(explainer.expected_value[0], shap_values[0], X_test, link="logit")

SHAP Interaction Values

SHAP interaction values are a generalization of SHAP values to higher order interactions. Fast exact computation of pairwise interactions are implemented for tree models with shap.TreeExplainer(model).shap_interaction_values(X). This returns a matrix for every prediction, where the main effects are on the diagonal and the interaction effects are off-diagonal. These values often reveal interesting hidden relationships, such as how the increased risk of death peaks for men at age 60 (see the NHANES notebook for details):

Sample notebooks

The notebooks below demonstrate different use cases for SHAP. Look inside the notebooks directory of the repository if you want to try playing with the original notebooks yourself.

TreeExplainer

An implementation of Tree SHAP, a fast and exact algorithm to compute SHAP values for trees and ensembles of trees.

DeepExplainer

An implementation of Deep SHAP, a faster (but only approximate) algorithm to compute SHAP values for deep learning models that is based on connections between SHAP and the DeepLIFT algorithm.

GradientExplainer

An implementation of expected gradients to approximate SHAP values for deep learning models. It is based on connections between SHAP and the Integrated Gradients algorithm. GradientExplainer is slower than DeepExplainer and makes different approximation assumptions.

LinearExplainer

For a linear model with independent features we can analytically compute the exact SHAP values. We can also account for feature correlation if we are willing to estimate the feature covariance matrix. LinearExplainer supports both of these options.

KernelExplainer

An implementation of Kernel SHAP, a model agnostic method to estimate SHAP values for any model. Because it makes no assumptions about the model type, KernelExplainer is slower than the other model type specific algorithms.

  • Census income classification with scikit-learn - Using the standard adult census income dataset, this notebook trains a k-nearest neighbors classifier using scikit-learn and then explains predictions using shap.

  • ImageNet VGG16 Model with Keras - Explain the classic VGG16 convolutional neural network's predictions for an image. This works by applying the model agnostic Kernel SHAP method to a super-pixel segmented image.

  • Iris classification - A basic demonstration using the popular iris species dataset. It explains predictions from six different models in scikit-learn using shap.

Documentation notebooks

These notebooks comprehensively demonstrate how to use specific functions and objects.

Methods Unified by SHAP

  1. LIME: Ribeiro, Marco Tulio, Sameer Singh, and Carlos Guestrin. "Why should i trust you?: Explaining the predictions of any classifier." Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, 2016.

  2. Shapley sampling values: Strumbelj, Erik, and Igor Kononenko. "Explaining prediction models and individual predictions with feature contributions." Knowledge and information systems 41.3 (2014): 647-665.

  3. DeepLIFT: Shrikumar, Avanti, Peyton Greenside, and Anshul Kundaje. "Learning important features through propagating activation differences." arXiv preprint arXiv:1704.02685 (2017).

  4. QII: Datta, Anupam, Shayak Sen, and Yair Zick. "Algorithmic transparency via quantitative input influence: Theory and experiments with learning systems." Security and Privacy (SP), 2016 IEEE Symposium on. IEEE, 2016.

  5. Layer-wise relevance propagation: Bach, Sebastian, et al. "On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation." PloS one 10.7 (2015): e0130140.

  6. Shapley regression values: Lipovetsky, Stan, and Michael Conklin. "Analysis of regression in game theory approach." Applied Stochastic Models in Business and Industry 17.4 (2001): 319-330.

  7. Tree interpreter: Saabas, Ando. Interpreting random forests. http://blog.datadive.net/interpreting-random-forests/

Citations

The algorithms and visualizations used in this package came primarily out of research in Su-In Lee's lab at the University of Washington, and Microsoft Research. If you use SHAP in your research we would appreciate a citation to the appropriate paper(s):

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

shap-0.46.0.tar.gz (1.2 MB view details)

Uploaded Source

Built Distributions

shap-0.46.0-cp312-cp312-win_amd64.whl (456.2 kB view details)

Uploaded CPython 3.12 Windows x86-64

shap-0.46.0-cp312-cp312-musllinux_1_2_x86_64.whl (1.5 MB view details)

Uploaded CPython 3.12 musllinux: musl 1.2+ x86-64

shap-0.46.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (540.7 kB view details)

Uploaded CPython 3.12 manylinux: glibc 2.17+ ARM64

shap-0.46.0-cp312-cp312-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (543.9 kB view details)

Uploaded CPython 3.12 manylinux: glibc 2.12+ x86-64 manylinux: glibc 2.17+ x86-64

shap-0.46.0-cp312-cp312-macosx_11_0_arm64.whl (455.3 kB view details)

Uploaded CPython 3.12 macOS 11.0+ ARM64

shap-0.46.0-cp312-cp312-macosx_10_9_x86_64.whl (459.3 kB view details)

Uploaded CPython 3.12 macOS 10.9+ x86-64

shap-0.46.0-cp311-cp311-win_amd64.whl (456.1 kB view details)

Uploaded CPython 3.11 Windows x86-64

shap-0.46.0-cp311-cp311-musllinux_1_2_x86_64.whl (1.5 MB view details)

Uploaded CPython 3.11 musllinux: musl 1.2+ x86-64

shap-0.46.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (537.8 kB view details)

Uploaded CPython 3.11 manylinux: glibc 2.17+ ARM64

shap-0.46.0-cp311-cp311-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (540.2 kB view details)

Uploaded CPython 3.11 manylinux: glibc 2.12+ x86-64 manylinux: glibc 2.17+ x86-64

shap-0.46.0-cp311-cp311-macosx_11_0_arm64.whl (455.8 kB view details)

Uploaded CPython 3.11 macOS 11.0+ ARM64

shap-0.46.0-cp311-cp311-macosx_10_9_x86_64.whl (459.3 kB view details)

Uploaded CPython 3.11 macOS 10.9+ x86-64

shap-0.46.0-cp310-cp310-win_amd64.whl (456.1 kB view details)

Uploaded CPython 3.10 Windows x86-64

shap-0.46.0-cp310-cp310-musllinux_1_2_x86_64.whl (1.5 MB view details)

Uploaded CPython 3.10 musllinux: musl 1.2+ x86-64

shap-0.46.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (537.8 kB view details)

Uploaded CPython 3.10 manylinux: glibc 2.17+ ARM64

shap-0.46.0-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (540.1 kB view details)

Uploaded CPython 3.10 manylinux: glibc 2.12+ x86-64 manylinux: glibc 2.17+ x86-64

shap-0.46.0-cp310-cp310-macosx_11_0_arm64.whl (455.8 kB view details)

Uploaded CPython 3.10 macOS 11.0+ ARM64

shap-0.46.0-cp310-cp310-macosx_10_9_x86_64.whl (459.3 kB view details)

Uploaded CPython 3.10 macOS 10.9+ x86-64

shap-0.46.0-cp39-cp39-win_amd64.whl (456.1 kB view details)

Uploaded CPython 3.9 Windows x86-64

shap-0.46.0-cp39-cp39-musllinux_1_2_x86_64.whl (1.5 MB view details)

Uploaded CPython 3.9 musllinux: musl 1.2+ x86-64

shap-0.46.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (537.7 kB view details)

Uploaded CPython 3.9 manylinux: glibc 2.17+ ARM64

shap-0.46.0-cp39-cp39-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (539.9 kB view details)

Uploaded CPython 3.9 manylinux: glibc 2.12+ x86-64 manylinux: glibc 2.17+ x86-64

shap-0.46.0-cp39-cp39-macosx_11_0_arm64.whl (455.8 kB view details)

Uploaded CPython 3.9 macOS 11.0+ ARM64

shap-0.46.0-cp39-cp39-macosx_10_9_x86_64.whl (459.3 kB view details)

Uploaded CPython 3.9 macOS 10.9+ x86-64

File details

Details for the file shap-0.46.0.tar.gz.

File metadata

  • Download URL: shap-0.46.0.tar.gz
  • Upload date:
  • Size: 1.2 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.12.4

File hashes

Hashes for shap-0.46.0.tar.gz
Algorithm Hash digest
SHA256 bdaa5b098be5a958348015e940f6fd264339b5db1e651f9898a3117be95b05a0
MD5 587f796fc2976830f52ec9e146df3369
BLAKE2b-256 47461b497452be642e19af56044814dfe32ee795805b443378821136729017a0

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp312-cp312-win_amd64.whl.

File metadata

  • Download URL: shap-0.46.0-cp312-cp312-win_amd64.whl
  • Upload date:
  • Size: 456.2 kB
  • Tags: CPython 3.12, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.12.4

File hashes

Hashes for shap-0.46.0-cp312-cp312-win_amd64.whl
Algorithm Hash digest
SHA256 0cbbf996537b2a42d3bc7f2a13492988822ee1bfd7220700989408dfb9e1c5ad
MD5 fdbc81515f6402ef484b4476dd9a36bb
BLAKE2b-256 0458b2ea558ec8d9ed3728e83dfacb1b920c54a1a1f6feee2632c04676c3c1e9

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp312-cp312-musllinux_1_2_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp312-cp312-musllinux_1_2_x86_64.whl
Algorithm Hash digest
SHA256 1230bf973463041dfa15734f290fbf3ab9c6e4e8222339c76f68fc355b940d80
MD5 08cff62e0c8a95132d1a707b4b8db22b
BLAKE2b-256 6eb6169de0d8971c91decd3dacfd63edeeedfc1bba30bfc6abf8480142aafd48

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl
Algorithm Hash digest
SHA256 b6e5dc5257b747a784f7a9b3acb64216a9011f01734f3c96b27fe5e15ae5f99f
MD5 cc3d66c5efe5752e7f77d1213147640c
BLAKE2b-256 8f8fca077689b76161b51b420031b88948ef92ade55730e85490215222734729

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp312-cp312-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp312-cp312-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 b216adf2a17b0e0694f17965ac29354ca8c4f27ac3c66f68bf6fc4cb2aa28207
MD5 08fc2ca74ab0f25496fedff00495a0ec
BLAKE2b-256 7f0ae3ab0dcddf4db1158fbf0d6c96348ba5f3031275f59088e0e3b7630cdcde

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp312-cp312-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp312-cp312-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 ab1fecfb43604605be17e26ae12bde4406c451c46b54b980d9570cec03fbc239
MD5 2a2fd5d3453e54a8f88a10d672b8f392
BLAKE2b-256 4d1ac00a1e7a68a4af29f2b40c8a8740dd241cba6cc58cd6ac266956a954a41d

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp312-cp312-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp312-cp312-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 a9cc9be191562bea1a782baff912854d267c6f4831bbf454d8d7bb7df7ddb214
MD5 28d02a9b985e0a769f5c3d011cfcb9cc
BLAKE2b-256 05c53c4fe600dd71fd2785d21f86a3e7f1f13de60c9b434052e05ba17598f81e

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp311-cp311-win_amd64.whl.

File metadata

  • Download URL: shap-0.46.0-cp311-cp311-win_amd64.whl
  • Upload date:
  • Size: 456.1 kB
  • Tags: CPython 3.11, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.12.4

File hashes

Hashes for shap-0.46.0-cp311-cp311-win_amd64.whl
Algorithm Hash digest
SHA256 c972a2efdc9fc00d543efaa55805eca947b8c418d065962d967824c2d5d295d0
MD5 4ddbf0188efb66605a7f9b7ffd12c4e2
BLAKE2b-256 8229923869e92c74bf07ec2b9a52ad5ac67d4184c873ba33ada7d4584356463a

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp311-cp311-musllinux_1_2_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp311-cp311-musllinux_1_2_x86_64.whl
Algorithm Hash digest
SHA256 943f0806fa00b4fafb174f172a73d88de2d8600e6d69c2e2bff833f00e6c4c21
MD5 959ff775055cf142ba6aa18743ede9f3
BLAKE2b-256 08e6027ca36efcc8871eda4084bde5e4658a90e84006086186e39588fd03b396

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl
Algorithm Hash digest
SHA256 70e06fdfdf53d5fb932c82f4529397552b262e0ccce734f5226fb1e1eab2bc3e
MD5 1753e12aec0aa4b0ca8a3e8f9ab1ab83
BLAKE2b-256 c374440eacbdf21c1b2e0a5b6962b79d4435e56a88588043d144a16c7785a596

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp311-cp311-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp311-cp311-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 13d36dc58d1e8c010feb4e7da71c77d23626a52d12d16b02869e793b11be4695
MD5 338ed2068921a5efcd6180608e4c56d2
BLAKE2b-256 066a09e3cb9864118337c0f3c2a0dc5add6b642e9f672665062e186d67ba992d

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp311-cp311-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp311-cp311-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 5bbdae4489577c6fce1cfe2d9d8f3d5b96d69284d29645fe651f78f6e965aeb4
MD5 2279e6f2776f9cbc382f29e742a66ee1
BLAKE2b-256 5f9edce41d5ec9e79add65faf4381d8d4492247b29daaa6cc7d7fd0298abc1e2

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp311-cp311-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp311-cp311-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 f18217c98f39fd485d541f6aab0b860b3be74b69b21d4faf11959e3fcba765c5
MD5 d338c979877d7b2a87b09be9d19b9100
BLAKE2b-256 e5a143bd69f32ddf381a09de18ea94d4b215d5ced3a24ff1a7b7d1a9401b5b85

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp310-cp310-win_amd64.whl.

File metadata

  • Download URL: shap-0.46.0-cp310-cp310-win_amd64.whl
  • Upload date:
  • Size: 456.1 kB
  • Tags: CPython 3.10, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.12.4

File hashes

Hashes for shap-0.46.0-cp310-cp310-win_amd64.whl
Algorithm Hash digest
SHA256 949bd7fa40371c3f1885a30ae0611dd481bf4ac90066ff726c73cb5bb393032b
MD5 2b52892a39eaff9578a234bc23cf4a21
BLAKE2b-256 aefef9e4d5e002bb58047c81edb6448579c179925c3807c98589ee70953587ab

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp310-cp310-musllinux_1_2_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp310-cp310-musllinux_1_2_x86_64.whl
Algorithm Hash digest
SHA256 0cf7c6e3f056cf3bfd16bcfd5744d0cc25b851555b1e750a3ab889b3077d2d05
MD5 60f39b69160fda78dd0430838258cecb
BLAKE2b-256 b4fcdd28e6838630cd436914116aa07a019753a40b956a05831b71bd3f7ce914

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl
Algorithm Hash digest
SHA256 c6097eb2ab7e8c194254bac3e462266490fbdd43bfe35a1014e9ee21c4ef10ee
MD5 0a55bdd7095b3ccb92408eb562db7683
BLAKE2b-256 351370e07364855b05d8aa628ec5aec4f038444ede0e26eee2be00c38077ee72

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 9633d3d7174acc01455538169ca6e6344f570530384548631aeadcf7bfdaaaea
MD5 d9dcecc3c0b4083986ba8fa210d04ba1
BLAKE2b-256 13a6b75760a52664dd82d530f9e232918bb74d1d6c39abcf34523c4f75cd4264

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp310-cp310-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp310-cp310-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 bccbb30ffbf8b9ed53e476d0c1319fdfcbeac455fe9df277fb0d570d92790e80
MD5 8dc887aee46c041c90d3b9e62530773f
BLAKE2b-256 00b32795a586a4446c8cbf04b6e8f15c19b4a6fb867e5c6cf9fcbca97d56a20b

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp310-cp310-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp310-cp310-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 905b2d7a0262ef820785a7c0e3c7f24c9d281e6f934edb65cbe811fe0e971187
MD5 15ebf4eb8b5e226a11ac254286e06f3a
BLAKE2b-256 13a897442ec8e7aaad01d860768232b3b7051adb0560a9c79e52ce5e1222cbf1

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp39-cp39-win_amd64.whl.

File metadata

  • Download URL: shap-0.46.0-cp39-cp39-win_amd64.whl
  • Upload date:
  • Size: 456.1 kB
  • Tags: CPython 3.9, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.12.4

File hashes

Hashes for shap-0.46.0-cp39-cp39-win_amd64.whl
Algorithm Hash digest
SHA256 b169b485a69f7d32e32fa64ad77be00129436c4455b9d0997b21b553f0becc8c
MD5 37bc7b7e6521cfd5f1397a98283a1e71
BLAKE2b-256 0248033ab9a2dee26d3de7e57cf532ab1d8408a608544c85ff98e6ea65775bdf

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp39-cp39-musllinux_1_2_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp39-cp39-musllinux_1_2_x86_64.whl
Algorithm Hash digest
SHA256 9f9f9727839e2459dfa4b4fbc190224e87f7b4b2a29f0e2a438500215921192b
MD5 cf8eb7cceb23fd97055f9fd20c7c266e
BLAKE2b-256 fa76d408da66d606e15d2ff6b6564d5023144679eebf4d6b298eb527ab56b043

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl
Algorithm Hash digest
SHA256 85a6ff9c9e15abd9a332360cff8d105165a600466167d6274dab468a050d005a
MD5 dbddad35f751a4979950a69d87391762
BLAKE2b-256 de90ffb554377d8b7ccecd0d40e56f9ecc9df0741eaf1f81a83ebb491f374307

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp39-cp39-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp39-cp39-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 99edc28daac4cbb98cd9f02febf4e9fbc6b9e3d24519c22ed59a98c68c47336c
MD5 083fe5641e1ec135bbf31fc7b6eff5eb
BLAKE2b-256 c65317cd5c57123f67b03c02d7a86ac5b9af76395f9e85e2e22960b259b2531a

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp39-cp39-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp39-cp39-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 0726f8c63f09dde586c9859ad315641f5a080e9aecf123a0cabc336b61703d66
MD5 a3151cf3e1b1805d89727029298bb709
BLAKE2b-256 66ffb8aaa11f1111a44c9e11bbb8302641106ca9a6531b3e3badb006b9dac5ad

See more details on using hashes here.

File details

Details for the file shap-0.46.0-cp39-cp39-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for shap-0.46.0-cp39-cp39-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 3c7d0c53a8cbefb2260ce28a98fa866c1a287770981f95c40a54f9d1082cbb31
MD5 f8c2a592a486632509c3e1d581c40218
BLAKE2b-256 64f9ddd643b2336139d9dab3a5facb356f5fd39aeefcf3333716bf7fc772319b

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page