HIVE: Evaluating the Human Interpretability of Visual Explanations

被引:38
|
作者
Kim, Sunnie S. Y. [1 ]
Meister, Nicole [1 ]
Ramaswamy, Vikram V. [1 ]
Fong, Ruth [1 ]
Russakovsky, Olga [1 ]
机构
[1] Princeton Univ, Princeton, NJ 08544 USA
来源
基金
美国国家科学基金会;
关键词
Interpretability; Explainable AI (XAI); Human studies; Evaluation framework; Human-centered AI;
D O I
10.1007/978-3-031-19775-8_17
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As AI technology is increasingly applied to high-impact, high-risk domains, there have been a number of new methods aimed at making AI models more human interpretable. Despite the recent growth of interpretability work, there is a lack of systematic evaluation of proposed techniques. In this work, we introduce HIVE (Human Interpretability of Visual Explanations), a novel human evaluation framework that assesses the utility of explanations to human users in AI-assisted decision making scenarios, and enables falsifiable hypothesis testing, cross-method comparison, and human-centered evaluation of visual interpretability methods. To the best of our knowledge, this is the first work of its kind. Using HIVE, we conduct IRB-approved human studies with nearly 1000 participants and evaluate four methods that represent the diversity of computer vision interpretability works: GradCAM, BagNet, ProtoPNet, and ProtoTree. Our results suggest that explanations engender human trust, even for incorrect predictions, yet are not distinct enough for users to distinguish between correct and incorrect predictions. We open-source HIVE to enable future studies and encourage more human-centered approaches to interpretability research. HIVE can be found at https://princetonvisualai.github.io/HIVE.
引用
收藏
页码:280 / 298
页数:19
相关论文
共 50 条
  • [41] Evaluating Recurrent Neural Network Explanations
    Arras, Leila
    Osman, Ahmed
    Mueller, Klaus-Robert
    Samek, Wojciech
    BLACKBOXNLP WORKSHOP ON ANALYZING AND INTERPRETING NEURAL NETWORKS FOR NLP AT ACL 2019, 2019, : 113 - 126
  • [42] Evaluating Alternative Explanations in Ecosystem Experiments
    Stephen R. Carpenter
    Jonathan J. Cole
    Timothy E. Essington
    James R. Hodgson
    Jeffrey N. Houser
    James F. Kitchell
    Michael L. Pace
    Ecosystems, 1998, 1 : 335 - 344
  • [43] Framework for Evaluating Faithfulness of Local Explanations
    Dasgupta, Sanjoy
    Frost, Nave
    Moshkovitz, Michal
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [44] Evaluating explanations of the Australian 'heroin shortage'
    Degenhardt, L
    Reuter, P
    Collins, L
    Hall, W
    ADDICTION, 2005, 100 (04) : 459 - 469
  • [45] Evaluating alternative explanations in ecosystem experiments
    Carpenter, SR
    Cole, JT
    Essington, TE
    Hodgson, JR
    Houser, JN
    Kitchell, JF
    Pace, ML
    ECOSYSTEMS, 1998, 1 (04) : 335 - 344
  • [46] Evaluating Causal Explanations of Specific Events
    Runde, Jochen
    de Rond, Mark
    ORGANIZATION STUDIES, 2010, 31 (04) : 431 - 450
  • [47] Evaluating the Performance of Interpretability Methods in Text Categorization Task
    Rogov, A. A.
    Loukachevitch, N. V.
    LOBACHEVSKII JOURNAL OF MATHEMATICS, 2024, 45 (03) : 1234 - 1245
  • [48] Visual recognition of honeybee behavior patterns at the hive entrance
    Sledevic, Tomyslav
    Serackis, Arturas
    Matuzevicius, Dalius
    Plonis, Darius
    Vdoviak, Gabriela
    PLOS ONE, 2025, 20 (02):
  • [49] FIND: A Function Description Benchmark for Evaluating Interpretability Methods
    Schwettmann, Sarah
    Shaham, Tamar Rott
    Materzynska, Joanna
    Chowdhury, Neil
    Li, Shuang
    Andreas, Jacob
    Bau, David
    Torralba, Antonio
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [50] Evaluating an Artistic Method for Depicting Human Visual Space
    Ruta, Nicole
    Burleigh, Alistair
    Vigars, Richard
    Barratt, Emma
    Peperell, Robert
    PERCEPTION, 2017, 46 (10) : 1209 - 1209