Interpreting Interpretability: Understanding Data Scientists' Use of Interpretability Tools for Machine Learning

被引:170
|
作者
Kaur, Harmanpreet [1 ]
Nori, Harsha [2 ]
Jenkins, Samuel [2 ]
Caruana, Rich [2 ]
Wallach, Hanna [2 ]
Vaughan, Jennifer Wortman [2 ]
机构
[1] Univ Michigan, Ann Arbor, MI 48109 USA
[2] Microsoft Res, Redmond, WA USA
关键词
interpretability; machine learning; user-centric evaluation; MODELS; CONTEXT;
D O I
10.1145/3313831.3376219
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Machine learning (ML) models are now routinely deployed in domains ranging from criminal justice to healthcare. With this newfound ubiquity, ML has moved beyond academia and grown into an engineering discipline. To that end, interpretability tools have been designed to help data scientists and machine learning practitioners better understand how ML models work. However, there has been little evaluation of the extent to which these tools achieve this goal. We study data scientists' use of two existing interpretability tools, the InterpretML implementation of GAMs and the SHAP Python package. We conduct a contextual inquiry (N=11) and a survey (N=197) of data scientists to observe how they use interpretability tools to uncover common issues that arise when building and evaluating ML models. Our results indicate that data scientists over-trust and misuse interpretability tools. Furthermore, few of our participants were able to accurately describe the visualizations output by these tools. We highlight qualitative themes for data scientists' mental models of interpretability tools. We conclude with implications for researchers and tool designers, and contextualize our findings in the social science literature.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Machine Learning Interpretability to Detect Fake Accounts in Instagram
    Sallah, Amine
    Alaoui, El Arbi Abdellaoui
    Agoujil, Said
    Nayyar, Anand
    [J]. INTERNATIONAL JOURNAL OF INFORMATION SECURITY AND PRIVACY, 2022, 16 (01)
  • [22] Philosophy of science at sea: Clarifying the interpretability of machine learning
    Beisbart, Claus
    Raz, Tim
    [J]. PHILOSOPHY COMPASS, 2022, 17 (06)
  • [23] Interpretability and Explainability of Machine Learning Models: Achievements and Challenges
    Henriques, J.
    Rocha, T.
    de Carvalho, P.
    Silva, C.
    Paredes, S.
    [J]. INTERNATIONAL CONFERENCE ON BIOMEDICAL AND HEALTH INFORMATICS 2022, ICBHI 2022, 2024, 108 : 81 - 94
  • [24] Interpretability of Machine Learning Solutions in Industrial Decision Engineering
    Kolyshkina, Inna
    Simoff, Simeon
    [J]. DATA MINING, AUSDM 2019, 2019, 1127 : 156 - 170
  • [25] Measuring Interpretability for Different Types of Machine Learning Models
    Zhou, Qing
    Liao, Fenglu
    Mou, Chao
    Wang, Ping
    [J]. TRENDS AND APPLICATIONS IN KNOWLEDGE DISCOVERY AND DATA MINING: PAKDD 2018 WORKSHOPS, 2018, 11154 : 295 - 308
  • [26] Interpretability and Fairness in Machine Learning: A Formal Methods Approach
    Ghosh, Bishwamittra
    [J]. PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 7083 - 7084
  • [27] Explainable AI: A Review of Machine Learning Interpretability Methods
    Linardatos, Pantelis
    Papastefanopoulos, Vasilis
    Kotsiantis, Sotiris
    [J]. ENTROPY, 2021, 23 (01) : 1 - 45
  • [28] The Importance of Interpretability and Validations of Machine-Learning Models
    Yamasawa, Daisuke
    Ozawa, Hideki
    Goto, Shinichi
    [J]. CIRCULATION JOURNAL, 2024, 88 (01) : 157 - 158
  • [29] Survey on Techniques, Applications and Security of Machine Learning Interpretability
    Ji, Shouling
    Li, Jinfeng
    Du, Tianyu
    Li, Bo
    [J]. Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2019, 56 (10): : 2071 - 2096
  • [30] Machine Learning Reimagined: The Promise of Interpretability to Combat Bias
    Maurer, Lydia R.
    Bertsimas, Dimitris
    Kaafarani, Haytham M. A.
    [J]. ANNALS OF SURGERY, 2022, 275 (06) : E738 - E739