Counterfactual explanations for misclassified images: How human and machine explanations differ

被引:3
|
作者
Delaney, Eoin [1 ,2 ,3 ]
Pakrashi, Arjun [1 ,3 ]
Greene, Derek [1 ,2 ,3 ]
Keane, Mark T. [1 ,3 ]
机构
[1] Univ Coll Dublin, Sch Comp Sci, Dublin, Ireland
[2] Insight Ctr Data Analyt, Dublin, Ireland
[3] VistaMilk SFI Res Ctr, Dublin, Ireland
基金
爱尔兰科学基金会;
关键词
XAI; Counterfactual explanation; User testing; BLACK-BOX;
D O I
10.1016/j.artint.2023.103995
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Counterfactual explanations have emerged as a popular solution for the eXplainable AI (XAI) problem of elucidating the predictions of black-box deep-learning systems because people easily understand them, they apply across different problem domains and seem to be legally compliant. Although over 100 counterfactual methods exist in the XAI literature, each claiming to generate plausible explanations akin to those preferred by people, few of these methods have actually been tested on users (similar to 7%). Even fewer studies adopt a user-centered perspective; for instance, asking people for their counterfactual explanations to determine their perspective on a "good explanation". This gap in the literature is addressed here using a novel methodology that (i) gathers human-generated counterfactual explanations for misclassified images, in two user studies and, then, (ii) compares these human-generated explanations to computationally-generated explanations for the same misclassifications. Results indicate that humans do not "minimally edit" images when generating counterfactual explanations. Instead, they make larger, "meaningful" edits that better approximate prototypes in the counterfactual class. An analysis based on "explanation goals" is proposed to account for this divergence between human and machine explanations. The implications of these proposals for future work are discussed. (c) 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY license (http://creativecommons .org /licenses /by /4 .0/).
引用
收藏
页数:25
相关论文
共 50 条
  • [1] Counterfactual Explanations for Misclassified Images: How Human and Machine Explanations Differ (Abstract Reprint)
    Delaney, Eoin
    Pakrashi, Arjun
    Greene, Derek
    Keane, Mark T.
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 20, 2024, : 22696 - 22696
  • [2] Counterfactual Explanations and Algorithmic Recourses for Machine Learning: A Review
    Verma, Sahil
    Boonsanong, Varich
    Hoang, Minh
    Hines, Keegan
    Dickerson, John
    Shah, Chirag
    ACM Computing Surveys, 2024, 56 (12)
  • [3] Counterfactual Visual Explanations
    Goyal, Yash
    Wu, Ziyan
    Ernst, Jan
    Batra, Dhruv
    Parikh, Devi
    Lee, Stefan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [4] The Skyline of Counterfactual Explanations for Machine Learning Decision Models
    Wang, Yongjie
    Ding, Qinxu
    Wang, Ke
    Liu, Yue
    Wu, Xingyu
    Wang, Jinglong
    Liu, Yong
    Miao, Chunyan
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 2030 - 2039
  • [5] Unjustified Classification Regions and Counterfactual Explanations in Machine Learning
    Laugel, Thibault
    Lesot, Marie-Jeanne
    Marsala, Christophe
    Renard, Xavier
    Detyniecki, Marcin
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT II, 2020, 11907 : 37 - 54
  • [6] ViCE: Visual Counterfactual Explanations for Machine Learning Models
    Gomez, Oscar
    Holter, Steffen
    Yuan, Jun
    Bertini, Enrico
    PROCEEDINGS OF THE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT USER INTERFACES, IUI 2020, 2020, : 531 - 535
  • [7] Beyond Trivial Counterfactual Explanations with Diverse Valuable Explanations
    Rodriguez, Pau
    Caccia, Massimo
    Lacoste, Alexandre
    Zamparo, Lee
    Laradji, Issam
    Charlin, Laurent
    Vazquez, David
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 1036 - 1045
  • [8] PreCoF: counterfactual explanations for fairness
    Sofie Goethals
    David Martens
    Toon Calders
    Machine Learning, 2024, 113 : 3111 - 3142
  • [9] On generating trustworthy counterfactual explanations
    Del Ser, Javier
    Barredo-Arrieta, Alejandro
    Diaz-Rodriguez, Natalia
    Herrera, Francisco
    Saranti, Anna
    Holzinger, Andreas
    INFORMATION SCIENCES, 2024, 655
  • [10] Explaining Machine Learning Classifiers through Diverse Counterfactual Explanations
    Mothilal, Ramaravind K.
    Sharma, Amit
    Tan, Chenhao
    FAT* '20: PROCEEDINGS OF THE 2020 CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, 2020, : 607 - 617