Explainable AI and Causal Understanding: Counterfactual Approaches Considered

被引:14
|
作者
Baron, Sam [1 ]
机构
[1] Australian Catholic Univ, Dianoia Inst Philosophy, 250 Victoria Parade, East Melbourne, Australia
关键词
Counterfactuals; Explanation; Causation; Interventions; Understanding; XAI;
D O I
10.1007/s11023-023-09637-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The counterfactual approach to explainable AI (XAI) seeks to provide understanding of AI systems through the provision of counterfactual explanations. In a recent systematic review, Chou et al. (Inform Fus 81:59-83, 2022) argue that the counterfactual approach does not clearly provide causal understanding. They diagnose the problem in terms of the underlying framework within which the counterfactual approach has been developed. To date, the counterfactual approach has not been developed in concert with the approach for specifying causes developed by Pearl (Causality: Models, reasoning, and inference. Cambridge University Press, 2000) and Woodward (Making things happen: A theory of causal explanation. Oxford University Press, 2003). In this paper, I build on Chou et al.'s work by applying the Pearl-Woodward approach. I argue that the standard counterfactual approach to XAI is capable of delivering causal understanding, but that there are limitations on its capacity to do so. I suggest a way to overcome these limitations.
引用
收藏
页码:347 / 377
页数:31
相关论文
共 50 条
  • [1] Explainable AI and Causal Understanding: Counterfactual Approaches Considered
    Sam Baron
    Minds and Machines, 2023, 33 : 347 - 377
  • [2] Counterfactual Explanations in Explainable AI: A Tutorial
    Wang, Cong
    Li, Xiao-Hui
    Han, Haocheng
    Wang, Shendi
    Wang, Luning
    Cao, Caleb Chen
    Chen, Lei
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 4080 - 4081
  • [3] The role of user feedback in enhancing understanding and trust in counterfactual explanations for explainable AI
    Suffian, Muhammad
    Kuhl, Ulrike
    Bogliolo, Alessandro
    Alonso-Moral, Jose M.
    INTERNATIONAL JOURNAL OF HUMAN-COMPUTER STUDIES, 2025, 199
  • [4] Understanding Interpretability: Explainable AI Approaches for Hate Speech Classifiers
    Yadav, Sargam
    Kaushik, Abhishek
    McDaid, Kevin
    EXPLAINABLE ARTIFICIAL INTELLIGENCE, XAI 2023, PT III, 2023, 1903 : 47 - 70
  • [5] FCE: Feedback Based Counterfactual Explanations for Explainable AI
    Suffian, Muhammad
    Graziani, Pierluigi
    Alonso, Jose M.
    Bogliolo, Alessandro
    IEEE ACCESS, 2022, 10 : 72363 - 72372
  • [6] FCE: Feedback Based Counterfactual Explanations for Explainable AI
    Suffian, Muhammad
    Graziani, Pierluigi
    Alonso, Jose M.
    Bogliolo, Alessandro
    IEEE Access, 2022, 10 : 72363 - 72372
  • [7] Understanding, Idealization, and Explainable AI
    Fleisher, Will
    EPISTEME-A JOURNAL OF INDIVIDUAL AND SOCIAL EPISTEMOLOGY, 2022, 19 (04): : 534 - 560
  • [8] Understanding AI: Explainable AI with interpretable KPI Labels
    Felix, Rudolf
    ATP MAGAZINE, 2020, (09): : 38 - 41
  • [9] Understanding the Limits of Explainable Ethical AI
    Peterson, Clayton
    Broersen, Jan
    INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS, 2024, 33 (03)
  • [10] XAIProcessLens: A Counterfactual-Based Dashboard for Explainable AI in Process Industries
    Manca, Gianluca
    Bhattacharya, Nilavra
    Maczey, Sylvia
    Ziobro, Dawid
    Brorsson, Emmanuel
    Bang, Magnus
    HHAI 2023: AUGMENTING HUMAN INTELLECT, 2023, 368 : 401 - 403