Improving the Interpretability of Deep Neural Networks with Knowledge Distillation

被引:70
|
作者
Liu, Xuan [1 ]
Wang, Xiaoguang [1 ,2 ]
Matwin, Stan [1 ,3 ]
机构
[1] Dalhousie Univ, Fac Comp Sci, Inst Big Data Analyt, Halifax, NS, Canada
[2] Alibaba Grp, Hangzhou, Zhejiang, Peoples R China
[3] Polish Acad Sci, Inst Comp Sci, Warsaw, Poland
关键词
interpretation; Neural Networks; Decision Tree; TensorFlow; dark knowledge; knowledge distillation;
D O I
10.1109/ICDMW.2018.00132
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Networks have achieved huge success at a wide spectrum of applications from language modeling, computer vision to speech recognition. However, nowadays, good performance alone is not enough to satisfy the needs of practical deployment where interpretability is demanded for cases involving ethics and mission critical applications. The complex models of Deep Neural Networks make it hard to understand and reason the predictions, which hinders its further progress. To tackle this problem, we apply the Knowledge Distillation technique to distill Deep Neural Networks into decision trees in order to attain good performance and interpretability simultaneously. We formulate the problem at hand as a multi-output regression problem and the experiments demonstrate that the student model achieves significantly better accuracy performance (about 1% to 5%) than vanilla decision trees at the same level of tree depth. The experiments are implemented on the TensorFlow platform to make it scalable to big datasets. To the best of our knowledge, we are the first to distill Deep Neural Networks into vanilla decision trees on multi-class datasets.
引用
收藏
页码:905 / 912
页数:8
相关论文
共 50 条
  • [1] IMPROVING THE INTERPRETABILITY OF DEEP NEURAL NETWORKS WITH STIMULATED LEARNING
    Tan, Shawn
    Sim, Khe Chai
    Gales, Mark
    [J]. 2015 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2015, : 617 - 623
  • [2] Improving Interpretability of Deep Neural Networks with Semantic Information
    Dong, Yinpeng
    Su, Hang
    Zhu, Jun
    Zhang, Bo
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 975 - 983
  • [3] Knowledge Distillation for Optimization of Quantized Deep Neural Networks
    Shin, Sungho
    Boo, Yoonho
    Sung, Wonyong
    [J]. 2020 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS), 2020, : 111 - 116
  • [4] Constructing Deep Spiking Neural Networks from Artificial Neural Networks with Knowledge Distillation
    Xu, Qi
    Li, Yaxin
    Shen, Jiangrong
    Liu, Jian K.
    Tang, Huajin
    Pan, Gang
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 7886 - 7895
  • [5] Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing Their Input Gradients
    Ros, Andrew Slavin
    Doshi-Velez, Finale
    [J]. THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 1660 - 1669
  • [6] Channel Planting for Deep Neural Networks using Knowledge Distillation
    Mitsuno, Kakeru
    Nomura, Yuichiro
    Kurita, Takio
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 7573 - 7579
  • [7] Soft Hybrid Knowledge Distillation against deep neural networks
    Zhang, Jian
    Tao, Ze
    Zhang, Shichao
    Qiao, Zike
    Guo, Kehua
    [J]. NEUROCOMPUTING, 2024, 570
  • [8] Improving the Interpretability of GradCAMs in Deep Classification Networks
    Schoettl, Alfred
    [J]. 3RD INTERNATIONAL CONFERENCE ON INDUSTRY 4.0 AND SMART MANUFACTURING, 2022, 200 : 620 - 628
  • [9] New Perspective of Interpretability of Deep Neural Networks
    Kimura, Masanari
    Tanaka, Masayuki
    [J]. 2020 3RD INTERNATIONAL CONFERENCE ON INFORMATION AND COMPUTER TECHNOLOGIES (ICICT 2020), 2020, : 78 - 85
  • [10] A Benchmark for Interpretability Methods in Deep Neural Networks
    Hooker, Sara
    Erhan, Dumitru
    Kindermans, Pieter-Jan
    Kim, Been
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32