HiBug: On Human-Interpretable Model Debug

被引:0
|
作者
Chen, Muxi [1 ]
Li, Yu [2 ]
Xu, Qiang [1 ]
机构
[1] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[2] Harbin Inst Technol, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Machine learning models can frequently produce systematic errors on critical subsets (or slices) of data that share common attributes. Discovering and explaining such model bugs is crucial for reliable model deployment. However, existing bug discovery and interpretation methods usually involve heavy human intervention and annotation, which can be cumbersome and have low bug coverage. In this paper, we propose HiBug, an automated framework for interpretable model debugging. Our approach utilizes large pre-trained models, such as chatGPT, to suggest human-understandable attributes that are related to the targeted computer vision tasks. By leveraging pre-trained vision-language models, we can efficiently identify common visual attributes of underperforming data slices using human-understandable terms. This enables us to uncover rare cases in the training data, identify spurious correlations in the model, and use the interpretable debug results to select or generate new training data for model improvement. Experimental results demonstrate the efficacy of the HiBug framework. Code is available at: https://github.com/cure- lab/HiBug.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Editorial: Human-Interpretable Machine Learning
    Tolomei, Gabriele
    Pinelli, Fabio
    Silvestri, Fabrizio
    FRONTIERS IN BIG DATA, 2022, 5
  • [2] Iris Recognition Based on Human-Interpretable Features
    Chen, Jianxu
    Shen, Feng
    Chen, Danny Z.
    Flynn, Patrick J.
    2015 IEEE INTERNATIONAL CONFERENCE ON IDENTITY, SECURITY AND BEHAVIOR ANALYSIS (ISBA), 2015,
  • [3] Iris Recognition Based on Human-Interpretable Features
    Chen, Jianxu
    Shen, Feng
    Chen, Danny Ziyi
    Flynn, Patrick J.
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2016, 11 (07) : 1476 - 1485
  • [4] Visual and Human-Interpretable Feedback for Assisting Physical Activity
    Antunes, Michel
    Baptista, Renato
    Demisse, Girum
    Aouada, Djamila
    Ottersten, Bjorn
    COMPUTER VISION - ECCV 2016 WORKSHOPS, PT II, 2016, 9914 : 115 - 129
  • [5] Human-Interpretable Diagnostic Information for Robotic Planning Systems
    Feng, Lu
    Humphrey, Laura
    Lee, Insup
    Topcu, Ufuk
    2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), 2016, : 1673 - 1680
  • [6] Human-interpretable and deep features for image privacy classification
    Baranouskaya, Darya
    Cavallaro, Andrea
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 3489 - 3492
  • [7] SENTECON: Leveraging Lexicons to Learn Human-Interpretable Language Representations
    Lin, Victoria
    Morency, Louis-Philippe
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 4312 - 4331
  • [8] GrASP: A Library for Extracting and Exploring Human-Interpretable Textual Patterns
    Lertvittayakumjorn, Piyawat
    Choshen, Leshem
    Shnarch, Eyal
    Toni, Francesca
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 6093 - 6103
  • [9] Generating Human-Interpretable Rules from Convolutional Neural Networks
    Pears, Russel
    Sharma, Ashwini Kumar
    Information (Switzerland), 2025, 16 (03)
  • [10] COMPREHENSIVE FACIAL EXPRESSION SYNTHESIS USING HUMAN-INTERPRETABLE LANGUAGE
    Hong, Joanna
    Kim, Jung Uk
    Lee, Sangmin
    Ro, Yong Man
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 1641 - 1645