Counterfactual explanations for misclassified images: How human and machine explanations differ

被引:3
|
作者
Delaney, Eoin [1 ,2 ,3 ]
Pakrashi, Arjun [1 ,3 ]
Greene, Derek [1 ,2 ,3 ]
Keane, Mark T. [1 ,3 ]
机构
[1] Univ Coll Dublin, Sch Comp Sci, Dublin, Ireland
[2] Insight Ctr Data Analyt, Dublin, Ireland
[3] VistaMilk SFI Res Ctr, Dublin, Ireland
基金
爱尔兰科学基金会;
关键词
XAI; Counterfactual explanation; User testing; BLACK-BOX;
D O I
10.1016/j.artint.2023.103995
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Counterfactual explanations have emerged as a popular solution for the eXplainable AI (XAI) problem of elucidating the predictions of black-box deep-learning systems because people easily understand them, they apply across different problem domains and seem to be legally compliant. Although over 100 counterfactual methods exist in the XAI literature, each claiming to generate plausible explanations akin to those preferred by people, few of these methods have actually been tested on users (similar to 7%). Even fewer studies adopt a user-centered perspective; for instance, asking people for their counterfactual explanations to determine their perspective on a "good explanation". This gap in the literature is addressed here using a novel methodology that (i) gathers human-generated counterfactual explanations for misclassified images, in two user studies and, then, (ii) compares these human-generated explanations to computationally-generated explanations for the same misclassifications. Results indicate that humans do not "minimally edit" images when generating counterfactual explanations. Instead, they make larger, "meaningful" edits that better approximate prototypes in the counterfactual class. An analysis based on "explanation goals" is proposed to account for this divergence between human and machine explanations. The implications of these proposals for future work are discussed. (c) 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY license (http://creativecommons .org /licenses /by /4 .0/).
引用
收藏
页数:25
相关论文
共 50 条
  • [21] Generating Robust Counterfactual Explanations
    Guyomard, Victor
    Fessant, Francoise
    Guyet, Thomas
    Bouadi, Tassadit
    Termier, Alexandre
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT III, 2023, 14171 : 394 - 409
  • [22] JUICE: JUstIfied Counterfactual Explanations
    Kuratomi, Alejandro
    Miliou, Ioanna
    Lee, Zed
    Lindgren, Tony
    Papapetrou, Panagiotis
    DISCOVERY SCIENCE (DS 2022), 2022, 13601 : 493 - 508
  • [23] Counterfactual explanations and how to find them: literature review and benchmarking
    Guidotti, Riccardo
    DATA MINING AND KNOWLEDGE DISCOVERY, 2024, 38 (05) : 2770 - 2824
  • [24] Disagreement amongst counterfactual explanations: how transparency can be misleading
    Brughmans, Dieter
    Melis, Lissa
    Martens, David
    TOP, 2024, 32 (3) : 429 - 462
  • [25] AdViCE: Aggregated Visual Counterfactual Explanations for Machine Learning Model Validation
    Gomez, Oscar
    Holter, Steffen
    Yuan, Jun
    Bertini, Enrico
    2021 IEEE VISUALIZATION CONFERENCE - SHORT PAPERS (VIS 2021), 2021, : 31 - 35
  • [26] The Robustness of Counterfactual Explanations Over Time
    Ferrario, Andrea
    Loi, Michele
    IEEE ACCESS, 2022, 10 : 82736 - 82750
  • [27] Ijuice: integer JUstIfied counterfactual explanations
    Kuratomi, Alejandro
    Miliou, Ioanna
    Lee, Zed
    Lindgren, Tony
    Papapetrou, Panagiotis
    MACHINE LEARNING, 2024, 113 (08) : 5731 - 5771
  • [28] Scaling Guarantees for Nearest Counterfactual Explanations
    Mohammadi, Kiarash
    Karimi, Amir-Hossein
    Barthe, Gilles
    Valera, Isabel
    AIES '21: PROCEEDINGS OF THE 2021 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2021, : 177 - 187
  • [29] Optimal Counterfactual Explanations in Tree Ensembles
    Parmentier, Axel
    Vida, Thibaut
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [30] STEEX: Steering Counterfactual Explanations with Semantics
    Jacob, Paul
    Zablocki, Eloi
    Ben-Younes, Hedi
    Chen, Mickael
    Perez, Patrick
    Cord, Matthieu
    COMPUTER VISION, ECCV 2022, PT XII, 2022, 13672 : 387 - 403