Interpretability, Then What? Editing Machine Learning Models to Reflect Human Knowledge and Values

被引:8
|
作者
Wang, Zijie J. [1 ]
Kale, Alex [2 ]
Nori, Harsha [3 ]
Stella, Peter [4 ]
Nunnally, Mark E. [4 ]
Chau, Duen Horng [1 ]
Vorvoreanu, Mihaela [3 ]
Vaughan, Jennifer Wortman [3 ]
Caruana, Rich [3 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
[2] Univ Washington, Seattle, WA 98195 USA
[3] Microsoft Res, New York, NY USA
[4] NYU, Langone Hlth, New York, NY 10003 USA
关键词
Interpretability; Model Editing; Accountability; Human Agency;
D O I
10.1145/3534678.3539074
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Machine learning (ML) interpretability techniques can reveal undesirable patterns in data that models exploit to make predictions-potentially causing harms once deployed. However, how to take action to address these patterns is not always clear. In a collaboration between ML and human-computer interaction researchers, physicians, and data scientists, we develop GAM Changer, the first interactive system to help domain experts and data scientists easily and responsibly edit Generalized Additive Models (GAMs) and fix problematic patterns. With novel interaction techniques, our tool puts interpretability into action-empowering users to analyze, validate, and align model behaviors with their knowledge and values. Physicians have started to use our tool to investigate and fix pneumonia and sepsis risk prediction models, and an evaluation with 7 data scientists working in diverse domains highlights that our tool is easy to use, meets their model editing needs, and fits into their current workflows. Built with modern web technologies, our tool runs locally in users' web browsers or computational notebooks, lowering the barrier to use. GAM Changer is available at the following public demo link: https://interpret.ml/gam-changer.
引用
收藏
页码:4132 / 4142
页数:11
相关论文
共 50 条
  • [1] Interpretability and Explainability of Machine Learning Models: Achievements and Challenges
    Henriques, J.
    Rocha, T.
    de Carvalho, P.
    Silva, C.
    Paredes, S.
    [J]. INTERNATIONAL CONFERENCE ON BIOMEDICAL AND HEALTH INFORMATICS 2022, ICBHI 2022, 2024, 108 : 81 - 94
  • [2] Measuring Interpretability for Different Types of Machine Learning Models
    Zhou, Qing
    Liao, Fenglu
    Mou, Chao
    Wang, Ping
    [J]. TRENDS AND APPLICATIONS IN KNOWLEDGE DISCOVERY AND DATA MINING: PAKDD 2018 WORKSHOPS, 2018, 11154 : 295 - 308
  • [3] The Importance of Interpretability and Validations of Machine-Learning Models
    Yamasawa, Daisuke
    Ozawa, Hideki
    Goto, Shinichi
    [J]. CIRCULATION JOURNAL, 2024, 88 (01) : 157 - 158
  • [4] Advancing interpretability of machine-learning prediction models
    Trenary, Laurie
    DelSole, Timothy
    [J]. ENVIRONMENTAL DATA SCIENCE, 2022, 1
  • [5] Accuracy, Fairness, and Interpretability of Machine Learning Criminal Recidivism Models
    Ingram, Eric
    Gursoy, Furkan
    Kakadiaris, Ioannis A.
    [J]. 2022 IEEE/ACM INTERNATIONAL CONFERENCE ON BIG DATA COMPUTING, APPLICATIONS AND TECHNOLOGIES, BDCAT, 2022, : 233 - 241
  • [6] Approach to provide interpretability in machine learning models for image classification
    Anja Stadlhofer
    Vitaliy Mezhuyev
    [J]. Industrial Artificial Intelligence, 1 (1):
  • [7] Applying Genetic Programming to Improve Interpretability in Machine Learning Models
    Ferreira, Leonardo Augusto
    Guimaraes, Frederico Gadelha
    Silva, Rodrigo
    [J]. 2020 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2020,
  • [8] Interpretability of machine learning-based prediction models in healthcare
    Stiglic, Gregor
    Kocbek, Primoz
    Fijacko, Nino
    Zitnik, Marinka
    Verbert, Katrien
    Cilar, Leona
    [J]. WILEY INTERDISCIPLINARY REVIEWS-DATA MINING AND KNOWLEDGE DISCOVERY, 2020, 10 (05)
  • [9] Analysis and interpretability of machine learning models to classify thyroid disease
    Akter, Sumya
    Mustafa, Hossen A.
    [J]. PLOS ONE, 2024, 19 (05):
  • [10] Wasserstein-based fairness interpretability framework for machine learning models
    Alexey Miroshnikov
    Konstandinos Kotsiopoulos
    Ryan Franks
    Arjun Ravi Kannan
    [J]. Machine Learning, 2022, 111 : 3307 - 3357