When are Post-hoc Conceptual Explanations Identifiable?

被引:0
|
作者
Leemann, Tobias [1 ,2 ]
Kirchhof, Michael [1 ]
Rong, Yao [1 ,2 ]
Kasneci, Enkelejda [2 ]
Kasneci, Gjergji [2 ]
机构
[1] Univ Tubingen, Tubingen, Germany
[2] Tech Univ Munich, Munich, Germany
来源
关键词
INDEPENDENT COMPONENT ANALYSIS; NONLINEAR ICA;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Interest in understanding and factorizing learned embedding spaces through conceptual explanations is steadily growing. When no human concept labels are available, concept discovery methods search trained embedding spaces for interpretable concepts like object shape or color that can provide post-hoc explanations for decisions. Unlike previous work, we argue that concept discovery should be identifiable, meaning that a number of known concepts can be provably recovered to guarantee reliability of the explanations. As a starting point, we explicitly make the connection between concept discovery and classical methods like Principal Component Analysis and Independent Component Analysis by showing that they can recover independent concepts under non-Gaussian distributions. For dependent concepts, we propose two novel approaches that exploit functional compositionality properties of image-generating processes. Our provably identifiable concept discovery methods substantially outperform competitors on a battery of experiments including hundreds of trained models and dependent concepts, where they exhibit up to 29% better alignment with the ground truth. Our results highlight the strict conditions under which reliable concept discovery without human labels can be guaranteed and provide a formal foundation for the domain. Our code is available online.
引用
收藏
页码:1207 / 1218
页数:12
相关论文
共 50 条
  • [11] Post-hoc recommendation explanations through an efficient exploitation of the DBpedia category hierarchy
    Du, Yu
    Ranwez, Sylvie
    Sutton-Charani, Nicolas
    Ranwez, Vincent
    KNOWLEDGE-BASED SYSTEMS, 2022, 245
  • [12] Ontology-Based Post-Hoc Explanations via Simultaneous Concept Extraction
    Ponomarev, Andrew
    Agafonov, Anton
    2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 887 - 890
  • [13] Evaluating Post-hoc Explanations for Graph Neural Networks via Robustness Analysis
    Fang, Junfeng
    Liu, Wei
    Gao, Yuan
    Liu, Zemin
    Zhang, An
    Wang, Xiang
    He, Xiangnan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [14] POST-HOC AND HYPOPROTHROMBINEMIA
    GALINSKY, RE
    FORNI, PJ
    MCGUIRE, GG
    TONG, TG
    BENOWITZ, N
    BECKER, CE
    ANNALS OF INTERNAL MEDICINE, 1975, 83 (02) : 286 - 286
  • [15] ON POST-HOC BLOCKING
    BONETT, DG
    EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 1982, 42 (01) : 35 - 39
  • [16] Assessing fidelity in XAI post-hoc techniques: A comparative study with ground truth explanations datasets
    Miro-Nicolau, Miquel
    Jaume-i-Capo, Antoni
    Moya-Alcover, Gabriel
    ARTIFICIAL INTELLIGENCE, 2024, 335
  • [17] A Quantitative Evaluation of Global, Rule-Based Explanations of Post-Hoc, Model Agnostic Methods
    Vilone, Giulia
    Longo, Luca
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2021, 4
  • [18] You said post-hoc?...
    de Roux-Serratrice, C
    Serratrice, J
    Champsaur, P
    Faucher, B
    Ené, N
    Granel, B
    Swiader, L
    Coulange, C
    Disdier, P
    Weiller, P
    REVUE DE MEDECINE INTERNE, 2005, 26 : S282 - S283
  • [19] Ontology-Based Post-Hoc Neural Network Explanations Via Simultaneous Concept Extraction
    Ponomarev, Andrew
    Agafonov, Anton
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 2, INTELLISYS 2023, 2024, 823 : 433 - 446
  • [20] How can I choose an explainer? An Application-grounded Evaluation of Post-hoc Explanations
    Jesus, Sergio
    Belem, Catarina
    Balayan, Vladimir
    Bento, Joao
    Saleiro, Pedro
    Bizarro, Pedro
    Gama, Joao
    PROCEEDINGS OF THE 2021 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2021, 2021, : 805 - 815