Explaining machine learning models using entropic variable projection

被引:1
|
作者
Bachoc, Francois [1 ]
Gamboa, Fabrice [1 ,2 ]
Halford, Max [3 ]
Loubes, Jean-Michel [1 ,2 ]
Risser, Laurent [1 ,2 ]
机构
[1] Inst Math Toulouse, Toulouse, France
[2] Artificial & Nat Intelligence Toulouse Inst 3IA AN, Toulouse, France
[3] Inst Rech Informat Toulouse, Toulouse, France
关键词
Explainability; Black-box decision rules; Kullback-Leibler divergence; Wasserstein distance;
D O I
10.1093/imaiai/iaad010
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
In this paper, we present a new explainability formalism designed to shed light on how each input variable of a test set impacts the predictions of machine learning models. Hence, we propose a group explainability formalism for trained machine learning decision rules, based on their response to the variability of the input variables distribution. In order to emphasize the impact of each input variable, this formalism uses an information theory framework that quantifies the influence of all input-output observations based on entropic projections. This is thus the first unified and model agnostic formalism enabling data scientists to interpret the dependence between the input variables, their impact on the prediction errors and their influence on the output predictions. Convergence rates of the entropic projections are provided in the large sample case. Most importantly, we prove that computing an explanation in our framework has a low algorithmic complexity, making it scalable to real-life large datasets. We illustrate our strategy by explaining complex decision rules learned using XGBoost, Random Forest or Deep Neural Network classifiers on various datasets such as Adult Income, MNIST, CelebA, Boston Housing, Iris, as well as synthetic ones. We finally make clear its differences with the explainability strategies LIME and SHAP, which are based on single observations. Results can be reproduced using the freely distributed Python toolbox .
引用
收藏
页数:30
相关论文
共 50 条
  • [21] Projection learning models
    Toh, Kar-Ann
    Teoh, Andrew Beng Jin
    ICIEA 2008: 3RD IEEE CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS, PROCEEDINGS, VOLS 1-3, 2008, : 2151 - 2155
  • [22] Sibyl: Explaining Machine Learning Models for High-Stakes Decision Making
    Zytek, Alexandra
    Liu, Dongyu
    Vaithianathan, Rhema
    Veeramachaneni, Kalyan
    EXTENDED ABSTRACTS OF THE 2021 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI'21), 2021,
  • [23] Exploring Gender Differences in Computational Thinking Learning in a VR Classroom: Developing Machine Learning Models Using Eye-Tracking Data and Explaining the Models
    Hong Gao
    Lisa Hasenbein
    Efe Bozkir
    Richard Göllner
    Enkelejda Kasneci
    International Journal of Artificial Intelligence in Education, 2023, 33 : 929 - 954
  • [24] Multi-criteria Approaches to Explaining Black Box Machine Learning Models
    Stefanowski, Jerzy
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS, ACIIDS 2023, PT II, 2023, 13996 : 195 - 208
  • [25] Interpretable machine learning models for predicting and explaining vehicle fuel consumption anomalies
    Barbado, Alberto
    Corcho, Oscar
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 115
  • [26] If Not Here, There. Explaining Machine Learning Models for Fault Localization in Optical Networks
    Karandin, Oleg
    Ayoub, Omran
    Musumeci, Francesco
    Hirota, Yusuke
    Awaji, Yoshinari
    Tornatore, Massimo
    2022 INTERNATIONAL CONFERENCE ON OPTICAL NETWORK DESIGN AND MODELLING (ONDM), 2022,
  • [27] Exploring Gender Differences in Computational Thinking Learning in a VR Classroom: Developing Machine Learning Models Using Eye-Tracking Data and Explaining the Models
    Gao, Hong
    Hasenbein, Lisa
    Bozkir, Efe
    Goellner, Richard
    Kasneci, Enkelejda
    INTERNATIONAL JOURNAL OF ARTIFICIAL INTELLIGENCE IN EDUCATION, 2023, 33 (04) : 929 - 954
  • [28] Explaining machine-learning models for gamma-ray detection and identification
    Bandstra, Mark S.
    Curtis, Joseph C.
    Ghawaly Jr, James M.
    Jones, A. Chandler
    Joshi, Tenzing H. Y.
    PLOS ONE, 2023, 18 (06):
  • [29] Large Language Models Are Latent Variable Models: Explaining and Finding Good Demonstrations for In-Context Learning
    Wang, Xinyi
    Zhu, Wanrong
    Saxon, Michael
    Steyvers, Mark
    Wang, William Yang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [30] vivid: An R package for Variable Importance and Variable Interactions Displays for Machine Learning Models
    Inglis, Alan
    Parnell, Andrew
    Hurley, Catherine
    R JOURNAL, 2023, 15 (02): : 344 - 361