Effect of Errors on the Evaluation of Machine Learning Systems

被引:1
|
作者
Bracamonte, Vanessa [1 ]
Hidano, Seira [1 ]
Kiyomoto, Shinsaku [1 ]
机构
[1] KDDI Res Inc, Saitama, Japan
关键词
User Perception; Errors; Machine Learning Model Evaluation; User Study; AUTOMATION; TRUST;
D O I
10.5220/0010839300003124
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Information such as accuracy and outcome explanations can be useful for the evaluation of machine learning systems, but they can also lead to over-trust. This means that an evaluator may not have suspicion that a machine learning system could have errors, and that they may overlook problems in the explanation of those systems. Research has shown that errors not only decrease trust but can also promote curiosity about the performance of the system. Therefore, presenting errors to evaluators may be an option to induce suspicion in the context of the evaluation of a machine learning system. In this paper, we evaluate this possibility by conducting three experiments where we asked participants to evaluate text classification systems. We presented two types of errors: incorrect predictions and errors in the explanation. The results show that patterns of errors in explanation negatively influenced willingness to recommend a system, and that fewer participants chose a system with higher accuracy when there was an error pattern, compared to when the errors were random. Moreover, more participants gave evidence from the explanations in their reason for their evaluation of the systems, suggesting that they were able to detect error patterns.
引用
收藏
页码:48 / 57
页数:10
相关论文
共 50 条
  • [21] Machine learning techniques for ocular errors analysis
    Libralao, GL
    de Almeida, OCP
    Netto, AV
    Delbem, ACB
    de Carvalho, ACPLF
    MACHINE LEARNING FOR SIGNAL PROCESSING XIV, 2004, : 569 - 578
  • [22] Machine Learning for Systems
    Litz, Heiner
    Hashemi, Milad
    IEEE MICRO, 2020, 40 (05) : 6 - 7
  • [23] Effect of errors on motor learning
    Maxwell, J
    Masters, R
    INTERNATIONAL JOURNAL OF PSYCHOLOGY, 2004, 39 (5-6) : 128 - 128
  • [24] Wizard of Errors: Introducing and Evaluating Machine Learning Errors in Wizard of Oz Studies
    Jansen, Anniek
    Colombo, Sara
    EXTENDED ABSTRACTS OF THE 2022 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2022, 2022,
  • [25] Boosting Recommendation Systems through an Offline Machine Learning Evaluation Approach
    Tsaku, Nelson Zange
    Kosaraju, Sai
    PROCEEDINGS OF THE 2019 ANNUAL ACM SOUTHEAST CONFERENCE (ACMSE 2019), 2019, : 182 - 185
  • [26] A machine learning methodology for reliability evaluation of complex chemical production systems
    Zhao, Fanrui
    Wu, Jinkui
    Zhao, Yuanpei
    Ji, Xu
    Zhou, Li
    Sun, Zhongping
    RSC ADVANCES, 2020, 10 (34) : 20374 - 20384
  • [27] Dataset Fault Tree Analysis for Systematic Evaluation of Machine Learning Systems
    Aoki, Toshiaki
    Kawakami, Daisuke
    Chida, Nobuo
    Tomita, Takashi
    2020 IEEE 25TH PACIFIC RIM INTERNATIONAL SYMPOSIUM ON DEPENDABLE COMPUTING (PRDC 2020), 2020, : 100 - 109
  • [28] On the Evaluation Measures for Machine Learning Algorithms for Safety-critical Systems
    Gharib, Mohamad
    Bondavalli, Andrea
    2019 15TH EUROPEAN DEPENDABLE COMPUTING CONFERENCE (EDCC 2019), 2019, : 141 - 144
  • [29] A machine learning-based usability evaluation method for eLearning systems
    Oztekin, Asil
    Delen, Dursun
    Turkyilmaz, Ali
    Zaim, Selim
    DECISION SUPPORT SYSTEMS, 2013, 56 : 63 - 73
  • [30] Machine Learning to Enhance Electronic Detection of Diagnostic Errors
    Zimolzak, Andrew J.
    Wei, Li
    Mir, Usman
    Gupta, Ashish
    Vaghani, Viralkumar
    Subramanian, Devika
    Singh, Hardeep
    JAMA NETWORK OPEN, 2024, 7 (09)