Regional Tree Regularization for Interpretability in Deep Neural Networks

被引:0
|
作者
Wu, Mike [1 ]
Parbhoo, Sonali [2 ,3 ]
Hughes, Michael C. [4 ]
Kindle, Ryan [5 ]
Celi, Leo [6 ]
Zazzi, Maurizio [7 ]
Roth, Volker [2 ]
Doshi-Velez, Finale [3 ]
机构
[1] Stanford Univ, Stanford, CA 94305 USA
[2] Univ Basel, Basel, Switzerland
[3] Harvard Univ, SEAS, Cambridge, MA 02138 USA
[4] Tufts Univ, Medford, MA 02155 USA
[5] Massachusetts Gen Hosp, Boston, MA 02114 USA
[6] MIT, Cambridge, MA 02139 USA
[7] Univ Siena, Siena, Italy
基金
瑞士国家科学基金会;
关键词
PREDICTION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The lack of interpretability remains a barrier to adopting deep neural networks across many safety-critical domains. Tree regularization was recently proposed to encourage a deep neural network's decisions to resemble those of a globally compact, axis-aligned decision tree. However, it is often unreasonable to expect a single tree to predict well across all possible inputs. In practice, doing so could lead to neither interpretable nor performant optima. To address this issue, we propose regional tree regularization - a method that encourages a deep model to be well-approximated by several separate decision trees specific to predefined regions of the input space. Across many datasets, including two healthcare applications, we show our approach delivers simpler explanations than other regularization schemes without compromising accuracy. Specifically, our regional regularizer finds many more "desirable" optima compared to global analogues.
引用
收藏
页码:6413 / 6421
页数:9
相关论文
共 50 条
  • [41] The Weights Reset Technique for Deep Neural Networks Implicit Regularization
    Plusch, Grigoriy
    Arsenyev-Obraztsov, Sergey
    Kochueva, Olga
    COMPUTATION, 2023, 11 (08)
  • [42] Deep Neural Networks Pruning via the Structured Perspective Regularization
    Cacciola, Matteo
    Frangioni, Antonio
    Li, Xinlin
    Lodi, Andrea
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2023, 5 (04): : 1051 - 1077
  • [43] The Analysis of Regularization in Deep Neural Networks Using Metagraph Approach
    Fedorenko, Yuriy S.
    Gapanyuk, Yuriy E.
    Minakova, Svetlana V.
    ADVANCES IN NEURAL COMPUTATION, MACHINE LEARNING, AND COGNITIVE RESEARCH, 2018, 736 : 3 - 8
  • [44] Hierarchical Group Sparse Regularization for Deep Convolutional Neural Networks
    Mitsuno, Kakeru
    Miyao, Junichi
    Kurita, Takio
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [45] Nonconvex Sparse Regularization for Deep Neural Networks and Its Optimality
    Ohn, Ilsang
    Kim, Yongdai
    NEURAL COMPUTATION, 2022, 34 (02) : 476 - 517
  • [46] Foothill: A Quasiconvex Regularization for Edge Computing of Deep Neural Networks
    Belbahri, Mouloud
    Sari, Eyyub
    Darabi, Sajad
    Nia, Vahid Partovi
    IMAGE ANALYSIS AND RECOGNITION (ICIAR 2019), PT II, 2019, 11663 : 3 - 14
  • [47] Interpretability-Guided Defense Against Backdoor Attacks to Deep Neural Networks
    Jiang, Wei
    Wen, Xiangyu
    Zhan, Jinyu
    Wang, Xupeng
    Song, Ziwei
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (08) : 2611 - 2624
  • [48] Transparency of deep neural networks for medical image analysis: A review of interpretability methods
    Salahuddin, Zohaib
    Woodruff, Henry C.
    Chatterjee, Avishek
    Lambin, Philippe
    COMPUTERS IN BIOLOGY AND MEDICINE, 2022, 140
  • [49] Reevaluating the Safety Impact of Inherent Interpretability on Deep Neural Networks for Pedestrian Detection
    Feifel, Patrick
    Bonarens, Frank
    Koester, Frank
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 29 - 37
  • [50] Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing Their Input Gradients
    Ros, Andrew Slavin
    Doshi-Velez, Finale
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 1660 - 1669