Biobjective gradient descent for feature selection on high dimension, low sample size data

被引:0
|
作者
Issa, Tina [1 ]
Angel, Eric [1 ]
Zehraoui, Farida [1 ]
机构
[1] Univ Paris Saclay, Univ Evry, IBISC, Evry Courcouronnes, France
来源
PLOS ONE | 2024年 / 19卷 / 07期
关键词
GENETIC ALGORITHM; OPTIMIZATION;
D O I
10.1371/journal.pone.0305654
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Even though deep learning shows impressive results in several applications, its use on problems with High Dimensions and Low Sample Size, such as diagnosing rare diseases, leads to overfitting. One solution often proposed is feature selection. In deep learning, along with feature selection, network sparsification is also used to improve the results when dealing with high dimensions low sample size data. However, most of the time, they are tackled as separate problems. This paper proposes a new approach that integrates feature selection, based on sparsification, into the training process of a deep neural network. This approach uses a constrained biobjective gradient descent method. It provides a set of Pareto optimal neural networks that make a trade-off between network sparsity and model accuracy. Results on both artificial and real datasets show that using a constrained biobjective gradient descent increases the network sparsity without degrading the classification performances. With the proposed approach, on an artificial dataset, the feature selection score reached 0.97 with a sparsity score of 0.92 with an accuracy of 0.9. For the same accuracy, none of the other methods reached a feature score above 0.20 and sparsity score of 0.35. Finally, statistical tests validate the results obtained on all datasets.
引用
收藏
页数:22
相关论文
共 50 条
  • [1] Ensemble feature selection in high dimension, low sample size datasets: Parallel and serial combination approaches
    Tsai, Chih-Fong
    Sung, Ya-Ting
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 203
  • [2] Analysis of feature selection stability on high dimension and small sample data
    Dernoncourt, David
    Hanczar, Blaise
    Zucker, Jean-Daniel
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2014, 71 : 681 - 693
  • [3] On Perfect Clustering of High Dimension, Low Sample Size Data
    Sarkar, Soham
    Ghosh, Anil K.
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (09) : 2257 - 2272
  • [4] Geometric representation of high dimension, low sample size data
    Hall, P
    Marron, JS
    Neeman, A
    [J]. JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2005, 67 : 427 - 444
  • [5] On feature selection protocols for very low-sample-size data
    Kuncheva, Ludmila I.
    Rodriguez, Juan J.
    [J]. PATTERN RECOGNITION, 2018, 81 : 660 - 673
  • [6] A variable selection method considering cluster loading for labeled high dimension low sample size data
    Chen, Jiaxin
    Sato-Ilic, Mika
    [J]. KNOWLEDGE-BASED AND INTELLIGENT INFORMATION & ENGINEERING SYSTEMS 19TH ANNUAL CONFERENCE, KES-2015, 2015, 60 : 850 - 859
  • [7] Classification for high-dimension low-sample size data
    Shen, Liran
    Er, Meng Joo
    Yin, Qingbo
    [J]. PATTERN RECOGNITION, 2022, 130
  • [8] Classification for high-dimension low-sample size data
    Shen, Liran
    Er, Meng Joo
    Yin, Qingbo
    [J]. PATTERN RECOGNITION, 2022, 130
  • [9] Deep Neural Networks for High Dimension, Low Sample Size Data
    Liu, Bo
    Wei, Ying
    Zhang, Yu
    Yang, Qiang
    [J]. PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2287 - 2293
  • [10] Improving the Performance of Feature Selection Methods with Low-Sample-Size Data
    Zheng, Wanwan
    Jin, Mingzhe
    [J]. COMPUTER JOURNAL, 2023, 66 (07): : 1664 - 1686