Improving the Interpretability of Deep Neural Networks with Knowledge Distillation

被引:70
|
作者
Liu, Xuan [1 ]
Wang, Xiaoguang [1 ,2 ]
Matwin, Stan [1 ,3 ]
机构
[1] Dalhousie Univ, Fac Comp Sci, Inst Big Data Analyt, Halifax, NS, Canada
[2] Alibaba Grp, Hangzhou, Zhejiang, Peoples R China
[3] Polish Acad Sci, Inst Comp Sci, Warsaw, Poland
关键词
interpretation; Neural Networks; Decision Tree; TensorFlow; dark knowledge; knowledge distillation;
D O I
10.1109/ICDMW.2018.00132
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Networks have achieved huge success at a wide spectrum of applications from language modeling, computer vision to speech recognition. However, nowadays, good performance alone is not enough to satisfy the needs of practical deployment where interpretability is demanded for cases involving ethics and mission critical applications. The complex models of Deep Neural Networks make it hard to understand and reason the predictions, which hinders its further progress. To tackle this problem, we apply the Knowledge Distillation technique to distill Deep Neural Networks into decision trees in order to attain good performance and interpretability simultaneously. We formulate the problem at hand as a multi-output regression problem and the experiments demonstrate that the student model achieves significantly better accuracy performance (about 1% to 5%) than vanilla decision trees at the same level of tree depth. The experiments are implemented on the TensorFlow platform to make it scalable to big datasets. To the best of our knowledge, we are the first to distill Deep Neural Networks into vanilla decision trees on multi-class datasets.
引用
收藏
页码:905 / 912
页数:8
相关论文
共 50 条
  • [31] A graph-based interpretability method for deep neural networks
    Wang, Tao
    Zheng, Xiangwei
    Zhang, Lifeng
    Cui, Zhen
    Xu, Chunyan
    [J]. NEUROCOMPUTING, 2023, 555
  • [32] Interpretability of deep neural networks: A review of methods, classification and hardware
    Antamis, Thanasis
    Drosou, Anastasis
    Vafeiadis, Thanasis
    Nizamis, Alexandros
    Ioannidis, Dimosthenis
    Tzovaras, Dimitrios
    [J]. NEUROCOMPUTING, 2024, 601
  • [33] Improving Deep Mutual Learning via Knowledge Distillation
    Lukman, Achmad
    Yang, Chuan-Kai
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (15):
  • [34] Improving Neural Topic Models with Wasserstein Knowledge Distillation
    Adhya, Suman
    Sanyal, Debarshi Kumar
    [J]. ADVANCES IN INFORMATION RETRIEVAL, ECIR 2023, PT II, 2023, 13981 : 321 - 330
  • [35] Improving Neural Topic Models using Knowledge Distillation
    Hoyle, Alexander
    Goel, Pranav
    Resnik, Philip
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1752 - 1771
  • [36] Some Shades of Grey! - Interpretability and Explainability of Deep Neural Networks
    Dengel, Andreas
    [J]. PROCEEDINGS OF THE ACM WORKSHOP ON CROSSMODAL LEARNING AND APPLICATION (WCRML'19), 2019, : 1 - 1
  • [37] Interpretability vs. Complexity: The Friction in Deep Neural Networks
    Amorim, Jose P.
    Abreu, Pedro H.
    Reyes, Mauricio
    Santos, Joao
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [38] Deep Convolutional Neural Networks Based on Knowledge Distillation for Offline Handwritten Chinese Character Recognition
    He, Hongli
    Zhu, Zongnan
    Li, Zhuo
    Dan, Yongping
    [J]. JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2024, 28 (02) : 231 - 238
  • [39] MULTI-TEACHER KNOWLEDGE DISTILLATION FOR COMPRESSED VIDEO ACTION RECOGNITION ON DEEP NEURAL NETWORKS
    Wu, Meng-Chieh
    Chiu, Ching-Te
    Wu, Kun-Hsuan
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 2202 - 2206
  • [40] Relevance aggregation for neural networks interpretability and knowledge discovery on tabular data
    Grisci, Bruno Iochins
    Krause, Mathias J.
    Dorn, Marcio
    [J]. INFORMATION SCIENCES, 2021, 559 : 111 - 129