Distribution-dependent feature selection for deep neural networks

被引:1
|
作者
Zhao, Xuebin [1 ]
Li, Weifu [1 ]
Chen, Hong [1 ]
Wang, Yingjie [2 ]
Chen, Yanhong [3 ]
John, Vijay [4 ]
机构
[1] Huazhong Agr Univ, Coll Sci, Wuhan 430062, Peoples R China
[2] Huazhong Agr Univ, Coll Informat, Wuhan 430062, Peoples R China
[3] Chinese Acad Sci, Natl Space Sci Ctr, Beijing 100190, Peoples R China
[4] Toyota Technol Inst, Res Ctr Smart Vehicles, Tempaku Ku, 2-12-1 Hisakata, Nagoya, Aichi 4688511, Japan
基金
中国国家自然科学基金;
关键词
Feature selection; Coronal mass ejections; Deep neural networks; Interpretability; Hypothesis-testing; FALSE DISCOVERY RATE; REGRESSION; FILTER;
D O I
10.1007/s10489-021-02663-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While deep neural networks (DNNs) have achieved impressive performance on a wide variety of tasks, the black-box nature hinders their applicability to high-risk, decision-making fields. In such fields, besides accurate prediction, it is also desired to provide interpretable insights into DNNs, e.g., screening important features based on their contributions to predictive accuracy. To improve the interpretability of DNNs, this paper originally proposes a new feature selection algorithm for DNNs by integrating the knockoff technique and the distribution information of irrelevant features. With the help of knockoff features and central limit theorem, we state that the irrelevant feature's statistic follows a known Gaussian distribution under mild conditions. This information is applied in hypothesis testing to discover key features associated with the DNNs. Empirical evaluations on simulated data demonstrate that the proposed method can select more true informative features with higher F-1 scores. Meanwhile, the Friedman test and the post-hoc Nemenyi test are employed to validate the superiority of the proposed method. Then we apply our method to Coronal Mass Ejections (CME) data and uncover the key features which contribute to the DNN-based CME arrival time.
引用
收藏
页码:4432 / 4442
页数:11
相关论文
共 50 条
  • [31] Feature selection using probabilistic neural networks
    Hunter, A
    NEURAL COMPUTING & APPLICATIONS, 2000, 9 (02): : 124 - 132
  • [32] Feature Selection Using Probabilistic Neural Networks
    A. Hunter
    Neural Computing & Applications, 2000, 9 : 124 - 132
  • [33] Distribution-dependent saccades in children with strabismus and in normals
    Kapoula, Z
    Bucci, MP
    EXPERIMENTAL BRAIN RESEARCH, 2002, 143 (02) : 264 - 268
  • [34] Distribution-Dependent PAC-Bayes Priors
    Lever, Guy
    Laviolette, Francois
    Shawe-Taylor, John
    ALGORITHMIC LEARNING THEORY, ALT 2010, 2010, 6331 : 119 - 133
  • [35] Comparison theorem for distribution-dependent neutral SFDEs
    Huang, Xing
    Yuan, Chenggui
    JOURNAL OF EVOLUTION EQUATIONS, 2021, 21 (01) : 653 - 670
  • [36] Comparison theorem for distribution-dependent neutral SFDEs
    Xing Huang
    Chenggui Yuan
    Journal of Evolution Equations, 2021, 21 : 653 - 670
  • [37] Generation and Analysis of Feature-Dependent Pseudo Noise for Training Deep Neural Networks
    Kamabattula, Sree Ram
    Musini, Kumudha
    Namazi, Babak
    Sankaranarayanan, Ganesh
    Devarajan, Venkat
    2021 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2021, : 2407 - 2413
  • [38] Distribution-Dependent Distance of First Two Moments
    Li, X. Rong
    2019 22ND INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION 2019), 2019,
  • [39] Distribution-dependent Vapnik-Chervonenkis bounds
    Vayatis, N
    Azencott, R
    COMPUTATIONAL LEARNING THEORY, 1999, 1572 : 230 - 240
  • [40] Improved feature processing for Deep Neural Networks
    Rath, Shakti P.
    Povey, Daniel
    Vesely, Karel
    Cernocky, Jan
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 109 - 113