HIVE: Evaluating the Human Interpretability of Visual Explanations

被引:38
|
作者
Kim, Sunnie S. Y. [1 ]
Meister, Nicole [1 ]
Ramaswamy, Vikram V. [1 ]
Fong, Ruth [1 ]
Russakovsky, Olga [1 ]
机构
[1] Princeton Univ, Princeton, NJ 08544 USA
来源
基金
美国国家科学基金会;
关键词
Interpretability; Explainable AI (XAI); Human studies; Evaluation framework; Human-centered AI;
D O I
10.1007/978-3-031-19775-8_17
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As AI technology is increasingly applied to high-impact, high-risk domains, there have been a number of new methods aimed at making AI models more human interpretable. Despite the recent growth of interpretability work, there is a lack of systematic evaluation of proposed techniques. In this work, we introduce HIVE (Human Interpretability of Visual Explanations), a novel human evaluation framework that assesses the utility of explanations to human users in AI-assisted decision making scenarios, and enables falsifiable hypothesis testing, cross-method comparison, and human-centered evaluation of visual interpretability methods. To the best of our knowledge, this is the first work of its kind. Using HIVE, we conduct IRB-approved human studies with nearly 1000 participants and evaluate four methods that represent the diversity of computer vision interpretability works: GradCAM, BagNet, ProtoPNet, and ProtoTree. Our results suggest that explanations engender human trust, even for incorrect predictions, yet are not distinct enough for users to distinguish between correct and incorrect predictions. We open-source HIVE to enable future studies and encourage more human-centered approaches to interpretability research. HIVE can be found at https://princetonvisualai.github.io/HIVE.
引用
收藏
页码:280 / 298
页数:19
相关论文
共 50 条
  • [21] Evaluating Self-attention Interpretability Through Human-Grounded Experimental Protocol
    Bhan, Milan
    Achache, Nina
    Legrand, Victor
    Blangero, Annabelle
    Chesneau, Nicolas
    EXPLAINABLE ARTIFICIAL INTELLIGENCE, XAI 2023, PT III, 2023, 1903 : 26 - 46
  • [22] Truthful meta-explanations for local interpretability of machine learning models
    Ioannis Mollas
    Nick Bassiliades
    Grigorios Tsoumakas
    Applied Intelligence, 2023, 53 : 26927 - 26948
  • [23] Truthful meta-explanations for local interpretability of machine learning models
    Mollas, Ioannis
    Bassiliades, Nick
    Tsoumakas, Grigorios
    APPLIED INTELLIGENCE, 2023, 53 (22) : 26927 - 26948
  • [24] PRINCE: Provider-side Interpretability with Counterfactual Explanations in Recommender Systems
    Ghazimatin, Azin
    Balalau, Oana
    Roy, Rishiraj Saha
    Weikum, Gerhard
    PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING (WSDM '20), 2020, : 196 - 204
  • [25] Beyond model interpretability: socio-structural explanations in machine learning
    Smart, Andrew
    Kasirzadeh, Atoosa
    AI & SOCIETY, 2024,
  • [26] Evaluating Robustness of Counterfactual Explanations
    Artelt, Andre
    Vaquet, Valerie
    Velioglu, Riza
    Hinder, Fabian
    Brinkrolf, Johannes
    Schilling, Malte
    Hammer, Barbara
    2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
  • [27] Evaluating Explanations by Cognitive Value
    Chander, Ajay
    Srinivasan, Ramya
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, CD-MAKE 2018, 2018, 11015 : 314 - 328
  • [28] Evaluating Attribution Methods in Machine Learning Interpretability
    Ratul, Qudrat E. Alahy
    Serra, Edoardo
    Cuzzocrea, Alfredo
    2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 5239 - 5245
  • [29] Evaluating the quality of visual explanations on chest X-ray images for thorax diseases classification
    Rahimiaghdam, Shakiba
    Alemdar, Hande
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (17): : 10239 - 10255
  • [30] Evaluating the Interpretability of Generative Models by Interactive Reconstruction
    Ross, Andrew
    Chen, Nina
    Hang, Elisa Zhao
    Glassman, Elena L.
    Doshi-Velez, Finale
    CHI '21: PROCEEDINGS OF THE 2021 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, 2021,