Interpreting Interpretability: Understanding Data Scientists' Use of Interpretability Tools for Machine Learning

被引:170
|
作者
Kaur, Harmanpreet [1 ]
Nori, Harsha [2 ]
Jenkins, Samuel [2 ]
Caruana, Rich [2 ]
Wallach, Hanna [2 ]
Vaughan, Jennifer Wortman [2 ]
机构
[1] Univ Michigan, Ann Arbor, MI 48109 USA
[2] Microsoft Res, Redmond, WA USA
关键词
interpretability; machine learning; user-centric evaluation; MODELS; CONTEXT;
D O I
10.1145/3313831.3376219
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Machine learning (ML) models are now routinely deployed in domains ranging from criminal justice to healthcare. With this newfound ubiquity, ML has moved beyond academia and grown into an engineering discipline. To that end, interpretability tools have been designed to help data scientists and machine learning practitioners better understand how ML models work. However, there has been little evaluation of the extent to which these tools achieve this goal. We study data scientists' use of two existing interpretability tools, the InterpretML implementation of GAMs and the SHAP Python package. We conduct a contextual inquiry (N=11) and a survey (N=197) of data scientists to observe how they use interpretability tools to uncover common issues that arise when building and evaluating ML models. Our results indicate that data scientists over-trust and misuse interpretability tools. Furthermore, few of our participants were able to accurately describe the visualizations output by these tools. We highlight qualitative themes for data scientists' mental models of interpretability tools. We conclude with implications for researchers and tool designers, and contextualize our findings in the social science literature.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Against Interpretability: a Critical Examination of the Interpretability Problem in Machine Learning
    Krishnan M.
    [J]. Philosophy & Technology, 2020, 33 (3) : 487 - 502
  • [2] Interpretability in HealthCare: A Comparative Study of Local Machine Learning Interpretability Techniques
    El Shawi, Radwa
    Sherif, Youssef
    Al-Mallah, Mouaz
    Sakr, Sherif
    [J]. 2019 IEEE 32ND INTERNATIONAL SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS (CBMS), 2019, : 275 - 280
  • [3] Interpretability in healthcare: A comparative study of local machine learning interpretability techniques
    ElShawi, Radwa
    Sherif, Youssef
    Al-Mallah, Mouaz
    Sakr, Sherif
    [J]. COMPUTATIONAL INTELLIGENCE, 2021, 37 (04) : 1633 - 1650
  • [4] A Study on Interpretability of Decision of Machine Learning
    Shirataki, Shohei
    Yamaguchi, Saneyasu
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2017, : 4830 - 4831
  • [5] Adding interpretability to predictive maintenance by machine learning on sensor data
    Steurtewagen, Bram
    Van den Poel, Dirk
    [J]. COMPUTERS & CHEMICAL ENGINEERING, 2021, 152
  • [6] A Review of Framework for Machine Learning Interpretability
    Araujo, Ivo de Abreu
    Torres, Renato Hidaka
    Sampaio Neto, Nelson Cruz
    [J]. AUGMENTED COGNITION, AC 2022, 2022, 13310 : 261 - 272
  • [7] Understanding gender differences in professional European football through machine learning interpretability and match actions data
    Garnica-Caparros, Marc
    Memmert, Daniel
    [J]. SCIENTIFIC REPORTS, 2021, 11 (01)
  • [8] Understanding gender differences in professional European football through machine learning interpretability and match actions data
    Marc Garnica-Caparrós
    Daniel Memmert
    [J]. Scientific Reports, 11
  • [9] Interpretability and Reproducability in Production Machine Learning Applications
    Ghanta, Sindhu
    Subramanian, Sriram
    Sundararaman, Swaminathan
    Khermosh, Lior
    Sridhar, Vinay
    Arteaga, Dulcardo
    Luo, Qianmei
    Das, Dhananjoy
    Talagala, Nisha
    [J]. 2018 17TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2018, : 658 - 664
  • [10] Evaluating Attribution Methods in Machine Learning Interpretability
    Ratul, Qudrat E. Alahy
    Serra, Edoardo
    Cuzzocrea, Alfredo
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 5239 - 5245