Measures for explainable AI: Explanation goodness, user satisfaction, mental models, curiosity, trust, and human-AI performance

被引:35
|
作者
Hoffman, Robert R. [1 ]
Mueller, Shane T. [2 ]
Klein, Gary [3 ]
Litman, Jordan [4 ]
机构
[1] Inst Human & Machine Cognit, Pensacola, FL 32502 USA
[2] Michigan Technol Univ, Dept Psychol, Houghton, MI USA
[3] MacroCognit LLC, Dayton, OH USA
[4] Univ Maine Machias, Dept Psychol, Machias, ME USA
来源
关键词
explanatory reasoning; machine-generated explanations; measurement; explanation goodness; mental models; trust; performance; SELF-EXPLANATION; AUTOMATION; CATEGORIZATION; INFORMATION; CALIBRATION; MACHINES; ILLUSION; SUPPORT; EXPERTS; LIMITS;
D O I
10.3389/fcomp.2023.1096257
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
If a user is presented an AI system that portends to explain how it works, how do we know whether the explanation works and the user has achieved a pragmatic understanding of the AI? This question entails some key concepts of measurement such as explanation goodness and trust. We present methods for enabling developers and researchers to: (1) Assess the a priori goodness of explanations, (2) Assess users' satisfaction with explanations, (3) Reveal user's mental model of an AI system, (4) Assess user's curiosity or need for explanations, (5) Assess whether the user's trust and reliance on the AI are appropriate, and finally, (6) Assess how the human-XAI work system performs. The methods we present derive from our integration of extensive research literatures and our own psychometric evaluations. We point to the previous research that led to the measurement scales which we aggregated and tailored specifically for the XAI context. Scales are presented in sufficient detail to enable their use by XAI researchers. For Mental Model assessment and Work System Performance, XAI researchers have choices. We point to a number of methods, expressed in terms of methods' strengths and weaknesses, and pertinent measurement issues.
引用
收藏
页数:15
相关论文
共 13 条
  • [1] Exploration of Explainable AI for Trust Development on Human-AI Interaction
    Bernardo, Ezekiel L.
    Seva, Rosemary R.
    PROCEEDINGS OF 2023 6TH ARTIFICIAL INTELLIGENCE AND CLOUD COMPUTING CONFERENCE, AICCC 2023, 2023, : 238 - 246
  • [2] The rationality of explanation or human capacity? Understanding the impact of explainable artificial intelligence on human-AI trust and decision performance
    Wang, Ping
    Ding, Heng
    INFORMATION PROCESSING & MANAGEMENT, 2024, 61 (04)
  • [3] Explainable AI improves task performance in human-AI collaboration
    Senoner, Julian
    Schallmoser, Simon
    Kratzwald, Bernhard
    Feuerriegel, Stefan
    Netland, Torbjorn
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [4] Trust in Human-AI Interaction: Scoping Out Models, Measures, and Methods
    Ueno, Takane
    Sawa, Yuto
    Kim, Yeongdae
    Urakami, Jacqueline
    Oura, Hiroki
    Seaborn, Katie
    EXTENDED ABSTRACTS OF THE 2022 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2022, 2022,
  • [5] Effects of Explanation Strategy and Autonomy of Explainable AI on Human-AI Collaborative Decision-making
    Wang, Bingcheng
    Yuan, Tianyi
    Rau, Pei-Luen Patrick
    INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2024, 16 (04) : 791 - 810
  • [6] Human-AI Collaboration: The Effect of AI Delegation on Human Task Performance and Task Satisfaction
    Hemmer, Patrick
    Westphal, Monika
    Schemmer, Max
    Vetter, Sebastian
    Vossing, Michael
    Satzger, Gerhard
    PROCEEDINGS OF 2023 28TH ANNUAL CONFERENCE ON INTELLIGENT USER INTERFACES, IUI 2023, 2023, : 453 - 463
  • [7] Trust in an AI versus a Human teammate: The effects of teammate identity and performance on Human-AI cooperation
    Zhang, Guanglu
    Chong, Leah
    Kotovsky, Kenneth
    Cagan, Jonathan
    COMPUTERS IN HUMAN BEHAVIOR, 2023, 139
  • [8] Towards Ethical AI: Empirically Investigating Dimensions of AI Ethics, Trust Repair, and Performance in Human-AI Teaming
    Schelble, Beau G.
    Lopez, Jeremy
    Textor, Claire
    Zhang, Rui
    McNeese, Nathan J.
    Pak, Richard
    Freeman, Guo
    HUMAN FACTORS, 2024, 66 (04) : 1037 - 1055
  • [9] Designing explainable AI to improve human-AI team performance: A medical stakeholder-driven scoping review
    Subramanian, Harishankar V.
    Canfield, Casey
    Shank, Daniel B.
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2024, 149
  • [10] The role of shared mental models in human-AI teams: a theoretical review
    Andrews, Robert W.
    Lilly, J. Mason
    Srivastava, Divya
    Feigh, Karen M.
    THEORETICAL ISSUES IN ERGONOMICS SCIENCE, 2023, 24 (02) : 129 - 175