A novel approach for generating explanations of the predictions of a generic ML model
Project description
Auracana XAI
Tree-based local explanations of machine learning model predictions. Implementation of the pipeline described in Parimbelli et al., 2021
About The Project
Increasingly complex learning methods such as boosting, bagging and deep learning have made ML models more accurate, but harder to understand and interpret. A tradeoff between performance and intelligibility is often to be faced, especially in high-stakes applications like medicine. This project propose a novel methodological approach for generating explanations of the predictions of a generic ML model, given a specific instance for which the prediction has been made, that can tackle both classification and regression tasks. Advantages of the proposed XAI approach include improved fidelity to the original model, the ability to deal with non-linear decision boundaries, and native support to both classification and regression problems.
Keywords: explainable AI, explanations, local explanation, fidelity, interpretability, transparency, trustworthy AI, black-box, machine learning, feature importance, decision tree, CART, AIM.
Paper
The araucanaxai package implements the pipeline described in Tree-based local explanations of machine learning model predictions - Parimbelli et al., 2021.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for araucanaxai-0.9.2-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | d8585cbbc7b37a7777e206a07361f6cb6ebcaa11367abc4ebd554aae937ee793 |
|
MD5 | 0b9a322db4853f26c8c777005ee00d60 |
|
BLAKE2b-256 | 4270cecb0119eb83c35f96425b46d864e8fb284797cc28810c463c915e0a3645 |