Ensembled sparse-input hierarchical networks for high-dimensional datasets

被引:4
|
作者
Feng, Jean [1 ]
Simon, Noah [2 ]
机构
[1] Univ Calif San Francisco, Dept Epidemiol & Biostat, San Francisco, CA 94143 USA
[2] Univ Washington, Dept Biostat, Seattle, WA 98195 USA
关键词
Bayesian model averaging; deep learning; grouping effect; lasso; network pruning; neural networks; VARIABLE SELECTION; REGULARIZATION;
D O I
10.1002/sam.11579
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In high-dimensional datasets where the number of covariates far exceeds the number of observations, the most popular prediction methods make strong modeling assumptions. Unfortunately, these methods struggle to scale up in model complexity as the number of observations grows. To this end, we consider using neural networks because they span a wide range of model capacities, from sparse linear models to deep neural networks. Because neural networks are notoriously tedious to tune and train, our aim is to develop a convenient procedure that employs a minimal number of hyperparameters. Our method, Ensemble by Averaging Sparse-Input hiERarchical networks (EASIER-net), employs only two L-1-penalty parameters, one that controls the input sparsity and another for the number of hidden layers and nodes. EASIER-net selects the true support with high probability when there is sufficient evidence; otherwise, it performs variable selection with uncertainty quantification, where strongly correlated covariates are selected at similar rates. On a large collection of gene expression datasets, EASIER-net achieved higher classification accuracy and selected fewer genes than existing methods. We found that EASIER-net adaptively selected the model complexity: it fit deep networks when there was sufficient information to learn nonlinearities and interactions and fit sparse logistic models for smaller datasets with less information.
引用
收藏
页码:736 / 750
页数:15
相关论文
共 50 条
  • [1] Online AUC Optimization for Sparse High-Dimensional Datasets
    Zhou, Baojian
    Ying, Yiming
    Skiena, Steven
    [J]. 20TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2020), 2020, : 881 - 890
  • [2] Sparse Bayesian hierarchical modeling of high-dimensional clustering problems
    Lian, Heng
    [J]. JOURNAL OF MULTIVARIATE ANALYSIS, 2010, 101 (07) : 1728 - 1737
  • [3] A hierarchical structure of extreme learning machine (HELM) for high-dimensional datasets with noise
    He, Yan-Lin
    Geng, Zhi-Qiang
    Xu, Yuan
    Zhu, Qun-Xiong
    [J]. NEUROCOMPUTING, 2014, 128 : 407 - 414
  • [4] On the challenges of learning with inference networks on sparse, high-dimensional data
    Krishnan, Rahul G.
    Liang, Dawen
    Hoffman, Matthew D.
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [5] High-dimensional sparse MANOVA
    Cai, T. Tony
    Xia, Yin
    [J]. JOURNAL OF MULTIVARIATE ANALYSIS, 2014, 131 : 174 - 196
  • [6] A method for learning a sparse classifier in the presence of missing data for high-dimensional biological datasets
    Severson, Kristen A.
    Monian, Brinda
    Love, J. Christopher
    Braatz, Richard D.
    [J]. BIOINFORMATICS, 2017, 33 (18) : 2897 - 2905
  • [7] Synthetic Generation of High-Dimensional Datasets
    Albuquerque, Georgia
    Loewe, Thomas
    Magnor, Marcus
    [J]. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2011, 17 (12) : 2317 - 2324
  • [8] Joining massive high-dimensional datasets
    Kahveci, T
    Lang, CA
    Singh, AK
    [J]. 19TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING, PROCEEDINGS, 2003, : 265 - 276
  • [9] Cluster validation for high-dimensional datasets
    Kim, M
    Yoo, H
    Ramakrishna, RS
    [J]. ARTIFICIAL INTELLIGENCE: METHODOLOGY, SYSTEMS, AND APPLICATIONS, PROCEEDINGS, 2004, 3192 : 178 - 187
  • [10] Distributed Learning of Deep Sparse Neural Networks for High-dimensional Classification
    Garg, Shweta
    Krishnan, R.
    Jagannathan, S.
    Samaranayake, V. A.
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2018, : 1587 - 1592