Deep Neural Networks for High Dimension, Low Sample Size Data

被引:0
|
作者
Liu, Bo [1 ]
Wei, Ying [1 ]
Zhang, Yu [1 ]
Yang, Qiang [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
FEATURE-SELECTION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNN) have achieved breakthroughs in applications with large sample size. However, when facing high dimension, low sample size (HDLSS) data, such as the phenotype prediction problem using genetic data in bioinformatics, DNN suffers from overfitting and high-variance gradients. In this paper, we propose a DNN model tailored for the HDLSS data, named Deep Neural Pursuit (DNP). DNP selects a subset of high dimensional features for the alleviation of overfitting and takes the average over multiple dropouts to calculate gradients with low variance. As the first DNN method applied on the HDLSS data, DNP enjoys the advantages of the high nonlinearity, the robustness to high dimensionality, the capability of learning from a small number of samples, the stability in feature selection, and the end-to-end training. We demonstrate these advantages of DNP via empirical results on both synthetic and real-world biological datasets.
引用
收藏
页码:2287 / 2293
页数:7
相关论文
共 50 条
  • [31] Intrinsic Dimensionality Estimation of High-Dimension, Low Sample Size Data with D-Asymptotics
    Yata, Kazuyoshi
    Aoshima, Makoto
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2010, 39 (8-9) : 1511 - 1521
  • [32] A variable selection method considering cluster loading for labeled high dimension low sample size data
    Chen, Jiaxin
    Sato-Ilic, Mika
    KNOWLEDGE-BASED AND INTELLIGENT INFORMATION & ENGINEERING SYSTEMS 19TH ANNUAL CONFERENCE, KES-2015, 2015, 60 : 850 - 859
  • [33] On asymptotic normality of cross data matrix-based PCA in high dimension low sample size
    Wang, Shao-Hsuan
    Huang, Su-Yun
    Chen, Ting-Li
    JOURNAL OF MULTIVARIATE ANALYSIS, 2020, 175
  • [34] Consistency of sparse PCA in High Dimension, Low Sample Size contexts
    Shen, Dan
    Shen, Haipeng
    Marron, J. S.
    JOURNAL OF MULTIVARIATE ANALYSIS, 2013, 115 : 317 - 333
  • [35] Data Dimension and Structure Effects in Predictive Performance of Deep Neural Networks
    Urda, Daniel
    Jerez, Jose M.
    Turias, Ignacio J.
    NEW TRENDS IN INTELLIGENT SOFTWARE METHODOLOGIES, TOOLS AND TECHNIQUES (SOMET_18), 2018, 303 : 361 - 372
  • [36] Boundary behavior in High Dimension, Low Sample Size asymptotics of PCA
    Jung, Sungkyu
    Sen, Arusharka
    Marron, J. S.
    JOURNAL OF MULTIVARIATE ANALYSIS, 2012, 109 : 190 - 203
  • [37] Design of input assignment and feedback gain for re-stabilizing undirected networks with High-Dimension Low-Sample-Size data
    Yasukata, Hitoshi
    Shen, Xun
    Sasahara, Hampei
    Imura, Jun-ichi
    Oku, Makito
    Aihara, Kazuyuki
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2023, 33 (12) : 6734 - 6753
  • [38] ON SIMULTANEOUS CALIBRATION OF TWO-SAMPLE t-TESTS FOR HIGH-DIMENSION LOW-SAMPLE-SIZE DATA
    Zhang, Chunming
    Jia, Shengji
    Wu, Yongfeng
    STATISTICA SINICA, 2021, 31 (03) : 1189 - 1214
  • [39] LOW: Training deep neural networks by learning optimal sample weights
    Santiago, Carlos
    Barata, Catarina
    Sasdelli, Michele
    Carneiro, Gustavo
    Nascimento, Jacinto C.
    PATTERN RECOGNITION, 2021, 110
  • [40] LOW: Training deep neural networks by learning optimal sample weights
    Santiago, Carlos
    Barata, Catarina
    Sasdelli, Michele
    Carneiro, Gustavo
    Nascimento, Jacinto C.
    Pattern Recognition, 2021, 110