Regional Tree Regularization for Interpretability in Deep Neural Networks

被引:0
|
作者
Wu, Mike [1 ]
Parbhoo, Sonali [2 ,3 ]
Hughes, Michael C. [4 ]
Kindle, Ryan [5 ]
Celi, Leo [6 ]
Zazzi, Maurizio [7 ]
Roth, Volker [2 ]
Doshi-Velez, Finale [3 ]
机构
[1] Stanford Univ, Stanford, CA 94305 USA
[2] Univ Basel, Basel, Switzerland
[3] Harvard Univ, SEAS, Cambridge, MA 02138 USA
[4] Tufts Univ, Medford, MA 02155 USA
[5] Massachusetts Gen Hosp, Boston, MA 02114 USA
[6] MIT, Cambridge, MA 02139 USA
[7] Univ Siena, Siena, Italy
基金
瑞士国家科学基金会;
关键词
PREDICTION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The lack of interpretability remains a barrier to adopting deep neural networks across many safety-critical domains. Tree regularization was recently proposed to encourage a deep neural network's decisions to resemble those of a globally compact, axis-aligned decision tree. However, it is often unreasonable to expect a single tree to predict well across all possible inputs. In practice, doing so could lead to neither interpretable nor performant optima. To address this issue, we propose regional tree regularization - a method that encourages a deep model to be well-approximated by several separate decision trees specific to predefined regions of the input space. Across many datasets, including two healthcare applications, we show our approach delivers simpler explanations than other regularization schemes without compromising accuracy. Specifically, our regional regularizer finds many more "desirable" optima compared to global analogues.
引用
收藏
页码:6413 / 6421
页数:9
相关论文
共 50 条
  • [21] Reliable interpretability of biology-inspired deep neural networks
    Wolfgang Esser-Skala
    Nikolaus Fortelny
    npj Systems Biology and Applications, 9
  • [22] A graph-based interpretability method for deep neural networks
    Wang, Tao
    Zheng, Xiangwei
    Zhang, Lifeng
    Cui, Zhen
    Xu, Chunyan
    NEUROCOMPUTING, 2023, 555
  • [23] Interpretability of deep neural networks: A review of methods, classification and hardware
    Antamis, Thanasis
    Drosou, Anastasis
    Vafeiadis, Thanasis
    Nizamis, Alexandros
    Ioannidis, Dimosthenis
    Tzovaras, Dimitrios
    NEUROCOMPUTING, 2024, 601
  • [24] A Graph-Based Interpretability Method for Deep Neural Networks
    Wang, Tao
    Zheng, Xiangwei
    Zhang, Lifeng
    Cui, Zhen
    Xu, Chunyan
    SSRN, 2022,
  • [25] Some Shades of Grey! - Interpretability and Explainability of Deep Neural Networks
    Dengel, Andreas
    PROCEEDINGS OF THE ACM WORKSHOP ON CROSSMODAL LEARNING AND APPLICATION (WCRML'19), 2019, : 1 - 1
  • [26] Interpretability vs. Complexity: The Friction in Deep Neural Networks
    Amorim, Jose P.
    Abreu, Pedro H.
    Reyes, Mauricio
    Santos, Joao
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [27] Deep neural networks regularization for structured output prediction
    Belharbi, Soufiane
    Herault, Romain
    Chatelain, Clement
    Adam, Sebastien
    NEUROCOMPUTING, 2018, 281 : 169 - 177
  • [28] Adaptive Knowledge Driven Regularization for Deep Neural Networks
    Luo, Zhaojing
    Cai, Shaofeng
    Cui, Can
    Ooi, Beng Chin
    Yang, Yang
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8810 - 8818
  • [29] Learning Credible Deep Neural Networks with Rationale Regularization
    Du, Mengnan
    Liu, Ninghao
    Yang, Fan
    Hu, Xia
    2019 19TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2019), 2019, : 150 - 159
  • [30] GradAug: A New Regularization Method for Deep Neural Networks
    Yang, Taojiannan
    Zhu, Sijie
    Chen, Chen
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33