Sparse Neural Additive Model: Interpretable Deep Learning with Feature Selection via Group Sparsity

被引:2
|
作者
Xu, Shiyun [1 ]
Bu, Zhiqi [1 ]
Chaudhari, Pratik [2 ]
Barnett, Ian J. [3 ]
机构
[1] Univ Penn, Dept Appl Math & Computat Sci, Philadelphia, PA 19104 USA
[2] Univ Penn, Dept Elect & Syst Engn, Philadelphia, PA 19104 USA
[3] Univ Penn, Dept Biostat Epidemiol & Informat, Philadelphia, PA 19104 USA
基金
美国国家科学基金会;
关键词
Interpretability; Additive Models; Group LASSO; Feature Selection; VARIABLE SELECTION; LASSO; REGRESSION; SHRINKAGE;
D O I
10.1007/978-3-031-43418-1_21
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Interpretable machine learning has demonstrated impressive performance while preserving explainability. In particular, neural additive models (NAM) offer the interpretability to the black-box deep learning and achieve state-of-the-art accuracy among the large family of generalized additive models. In order to empower NAM with feature selection and improve the generalization, we propose the sparse neural additive models (SNAM) that employ the group sparsity regularization (e.g. Group LASSO), where each feature is learned by a sub-network whose trainable parameters are clustered as a group. We study the theoretical properties for SNAM with novel techniques to tackle the non-parametric truth, thus extending from classical sparse linear models such as the LASSO, which only works on the parametric truth. Specifically, we show that SNAM with subgradient and proximal gradient descents provably converges to zero training loss as t -> infinity, and that the estimation error of SNAM vanishes asymptotically as n -> infinity. We also prove that SNAM, similar to LASSO, can have exact support recovery, i.e. perfect feature selection, with appropriate regularization. Moreover, we show that the SNAM can generalize well and preserve the 'identifiability', recovering each feature's effect. We validate our theories via extensive experiments and further testify to the good accuracy and efficiency of SNAM (Appendix can be found at https://arxiv.org/abs/2202.12482.).
引用
收藏
页码:343 / 359
页数:17
相关论文
共 50 条
  • [1] A Multiobjective Sparse Feature Learning Model for Deep Neural Networks
    Gong, Maoguo
    Liu, Jia
    Li, Hao
    Cai, Qing
    Su, Linzhi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (12) : 3263 - 3277
  • [2] Unsupervised feature selection via joint local learning and group sparse regression
    Yue WU
    Can WANG
    Yue-qing ZHANG
    Jia-jun BU
    Frontiers of Information Technology & Electronic Engineering, 2019, 20 (04) : 538 - 553
  • [3] Unsupervised feature selection via joint local learning and group sparse regression
    Yue Wu
    Can Wang
    Yue-qing Zhang
    Jia-jun Bu
    Frontiers of Information Technology & Electronic Engineering, 2019, 20 : 538 - 553
  • [4] Unsupervised feature selection via joint local learning and group sparse regression
    Wu, Yue
    Wang, Can
    Zhang, Yue-qing
    Bu, Jia-jun
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2019, 20 (04) : 538 - 553
  • [5] Sparse Interaction Additive Networks via Feature Interaction Detection and Sparse Selection
    Enouen, James
    Liu, Yan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [6] Towards Interpretable Deep Learning: A Feature Selection Framework for Prognostics and Health Management Using Deep Neural Networks
    Barraza, Joaquin Figueroa
    Droguett, Enrique Lopez
    Martins, Marcelo Ramos
    SENSORS, 2021, 21 (17)
  • [7] Heterogeneous Feature Selection With Multi-Modal Deep Neural Networks and Sparse Group LASSO
    Zhao, Lei
    Hu, Qinghua
    Wang, Wenwu
    IEEE TRANSACTIONS ON MULTIMEDIA, 2015, 17 (11) : 1936 - 1948
  • [8] Automated Feature Document Review via Interpretable Deep Learning
    Ye, Ming
    Chen, Yuanfan
    Zhang, Xin
    He, Jinning
    Cao, Jicheng
    Liu, Dong
    Gao, Jing
    Dai, Hailiang
    Cheng, Shengyu
    2023 IEEE/ACM 45TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING: COMPANION PROCEEDINGS, ICSE-COMPANION, 2023, : 351 - 354
  • [9] Group variable selection via group sparse neural network
    Zhang, Xin
    Zhao, Junlong
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2024, 192
  • [10] Is Interpretable Machine Learning Effective at Feature Selection for Neural Learning-to-Rank?
    Lyu, Lijun
    Roy, Nirmal
    Oosterhuis, Harrie
    Anand, Avishek
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT IV, 2024, 14611 : 384 - 402