HIVE: Evaluating the Human Interpretability of Visual Explanations

被引:38
|
作者
Kim, Sunnie S. Y. [1 ]
Meister, Nicole [1 ]
Ramaswamy, Vikram V. [1 ]
Fong, Ruth [1 ]
Russakovsky, Olga [1 ]
机构
[1] Princeton Univ, Princeton, NJ 08544 USA
来源
基金
美国国家科学基金会;
关键词
Interpretability; Explainable AI (XAI); Human studies; Evaluation framework; Human-centered AI;
D O I
10.1007/978-3-031-19775-8_17
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As AI technology is increasingly applied to high-impact, high-risk domains, there have been a number of new methods aimed at making AI models more human interpretable. Despite the recent growth of interpretability work, there is a lack of systematic evaluation of proposed techniques. In this work, we introduce HIVE (Human Interpretability of Visual Explanations), a novel human evaluation framework that assesses the utility of explanations to human users in AI-assisted decision making scenarios, and enables falsifiable hypothesis testing, cross-method comparison, and human-centered evaluation of visual interpretability methods. To the best of our knowledge, this is the first work of its kind. Using HIVE, we conduct IRB-approved human studies with nearly 1000 participants and evaluate four methods that represent the diversity of computer vision interpretability works: GradCAM, BagNet, ProtoPNet, and ProtoTree. Our results suggest that explanations engender human trust, even for incorrect predictions, yet are not distinct enough for users to distinguish between correct and incorrect predictions. We open-source HIVE to enable future studies and encourage more human-centered approaches to interpretability research. HIVE can be found at https://princetonvisualai.github.io/HIVE.
引用
收藏
页码:280 / 298
页数:19
相关论文
共 50 条
  • [1] Interpretability Benchmark for Evaluating Spatial Misalignment of Prototypical Parts Explanations
    Sacha, Mikolaj
    Jura, Bartosz
    Rymarczyk, Dawid
    Struski, Lukasz
    Tabor, Jacek
    Zielinski, Bartosz
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 19, 2024, : 21563 - 21573
  • [2] Contrastive Explanations for Model Interpretability
    Jacovi, Alon
    Swayamdipta, Swabha
    Ravfogel, Shauli
    Elazar, Yanai
    Choi, Yejin
    Goldberg, Yoav
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1597 - 1611
  • [3] Privacy-Preserving Case-Based Explanations: Enabling Visual Interpretability by Protecting Privacy
    Montenegro, Helena
    Silva, Wilson
    Gaudio, Alex
    Fredrikson, Matt
    Smailagic, Asim
    Cardoso, Jaime S.
    IEEE ACCESS, 2022, 10 : 28333 - 28347
  • [4] A Protocol for Evaluating Model Interpretation Methods from Visual Explanations
    Behzadi-Khormouji, Hamed
    Oramas, Jose
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 1421 - 1429
  • [5] On the Granularity of Explanations in Model Agnostic NLP Interpretability
    Rychener, Yves
    Renard, Xavier
    Seddah, Djame
    Frossard, Pascal
    Detyniecki, Marcin
    MACHINE LEARNING AND PRINCIPLES AND PRACTICE OF KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT I, 2023, 1752 : 498 - 512
  • [6] Explaining Explanations: An Overview of Interpretability of Machine Learning
    Gilpin, Leilani H.
    Bau, David
    Yuan, Ben Z.
    Bajwa, Ayesha
    Specter, Michael
    Kagal, Lalana
    2018 IEEE 5TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2018, : 80 - 89
  • [7] Teaching Categories to Human Learners with Visual Explanations
    Mac Aodha, Oisin
    Su, Shihan
    Chen, Yuxin
    Perona, Pietro
    Yue, Yisong
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 3820 - 3828
  • [8] Evaluating Quality of Visual Explanations of Deep Learning Models for Vision Tasks
    Yang, Yuqing
    Mahmoudpour, Saeed
    Schelkens, Peter
    Deligiannis, Nikos
    2023 15TH INTERNATIONAL CONFERENCE ON QUALITY OF MULTIMEDIA EXPERIENCE, QOMEX, 2023, : 159 - 164
  • [9] Leveraging saliency priors and explanations for enhanced consistent interpretability
    Dong, Liang
    Chen, Leiyang
    Fu, Zhongwang
    Zheng, Chengliang
    Cui, Xiaohui
    Shen, Zhidong
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 249
  • [10] Evaluating the Interpretability of Threshold Operators
    Righetti, Guendalina
    Porello, Daniele
    Confalonieri, Roberto
    KNOWLEDGE ENGINEERING AND KNOWLEDGE MANAGEMENT, EKAW 2022, 2022, 13514 : 136 - 151