CDeepEx: Contrastive Deep Explanations

被引:1
|
作者
Feghahati, Amir [1 ]
Shelton, Christian R. [1 ]
Pazzani, Michael J. [1 ]
Tang, Kevin [1 ]
机构
[1] Univ Calif Riverside, Riverside, CA 92521 USA
基金
美国国家科学基金会;
关键词
D O I
10.3233/FAIA200212
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a method which can visually explain the classification decision of deep neural networks (DNNs). Many methods have been proposed in machine learning and computer vision seeking to clarify the decision of machine learning black boxes, specifically DNNs. All of these methods try to gain insight into why the network "chose class A" as an answer. Humans search for explanations by asking two types of questions. The first question is, "Why did you choose this answer?" The second question asks, "Why did you not choose answer B over A?" The previously proposed methods are not able to provide the latter directly or efficiently. We introduce a method capable of answering the second question both directly and efficiently. In this work, we limit the inputs to be images. In general, the proposed method generates explanations in the input space of any model capable of efficient evaluation and gradient evaluation. It does not require any knowledge of the underlying classifier nor use heuristics in its explanation generation, and it is computationally fast to evaluate. We provide extensive experimental results on three different datasets, showing the robustness of our approach, and its superiority for gaining insight into the inner representations of machine learning models. As an example, we demonstrate our method can detect and explain how a network trained to recognize hair color actually detects eye color, whereas other methods cannot find this bias in the trained classifier.
引用
收藏
页码:1143 / 1151
页数:9
相关论文
共 50 条
  • [1] The demand for contrastive explanations
    Nadine Elzein
    [J]. Philosophical Studies, 2019, 176 : 1325 - 1339
  • [2] The demand for contrastive explanations
    Elzein, Nadine
    [J]. PHILOSOPHICAL STUDIES, 2019, 176 (05) : 1325 - 1339
  • [3] Contrastive Explanations for a Deep Learning Model on Time-Series Data
    Labaien, Jokin
    Zugasti, Ekhi
    De Carlos, Xabier
    [J]. BIG DATA ANALYTICS AND KNOWLEDGE DISCOVERY (DAWAK 2020), 2020, 12393 : 235 - 244
  • [4] Contrastive explanations: A dilemma for libertarians
    Levy, N
    [J]. DIALECTICA, 2005, 59 (01) : 51 - 61
  • [5] A characterization of contrastive explanations computation
    Kean, A
    [J]. PRICAI'98: TOPICS IN ARTIFICIAL INTELLIGENCE, 1998, 1531 : 599 - 610
  • [6] Efficient computation of contrastive explanations
    Artelt, Andre
    Hammer, Barbara
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [7] Consistent Explanations by Contrastive Learning
    Pillai, Vipin
    Koohpayegani, Soroush Abbasi
    Ouligian, Ashley
    Fong, Dennis
    Pirsiavash, Hamed
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 10203 - 10212
  • [8] CONTRASTIVE EXPLANATIONS IN EVOLUTIONARY BIOLOGY
    Boulter, Stephen
    [J]. RATIO, 2012, 25 (04) : 425 - 441
  • [9] CONTRASTIVE EXPLANATIONS IN NEURAL NETWORKS
    Prabhushankar, Mohit
    Kwon, Gukyeong
    Temel, Dogancan
    AlRegib, Ghassan
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 3289 - 3293
  • [10] Contrastive Explanations for Model Interpretability
    Jacovi, Alon
    Swayamdipta, Swabha
    Ravfogel, Shauli
    Elazar, Yanai
    Choi, Yejin
    Goldberg, Yoav
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1597 - 1611