A hybrid feature selection algorithm combining information gain and grouping particle swarm optimization for cancer diagnosis

被引:0
|
作者
Yang, Fangyuan [1 ]
Xu, Zhaozhao [2 ]
Wang, Hong [1 ]
Sun, Lisha [1 ]
Zhai, Mengjiao [1 ]
Zhang, Juan [1 ]
机构
[1] Henan Polytech Univ, Affiliated Hosp 1, Dept Gynecol Oncol, Jiaozuo, Henan, Peoples R China
[2] Henan Polytech Univ, Sch Comp Sci & Technol, Jiaozuo, Henan, Peoples R China
来源
PLOS ONE | 2024年 / 19卷 / 03期
关键词
WRAPPER FEATURE-SELECTION; CLASSIFICATION; FILTER;
D O I
10.1371/journal.pone.0290332
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Background Cancer diagnosis based on machine learning has become a popular application direction. Support vector machine (SVM), as a classical machine learning algorithm, has been widely used in cancer diagnosis because of its advantages in high-dimensional and small sample data. However, due to the high-dimensional feature space and high feature redundancy of gene expression data, SVM faces the problem of poor classification effect when dealing with such data.Methods Based on this, this paper proposes a hybrid feature selection algorithm combining information gain and grouping particle swarm optimization (IG-GPSO). The algorithm firstly calculates the information gain values of the features and ranks them in descending order according to the value. Then, ranked features are grouped according to the information index, so that the features in the group are close, and the features outside the group are sparse. Finally, grouped features are searched using grouping PSO and evaluated according to in-group and out-group.Results Experimental results show that the average accuracy (ACC) of the SVM on the feature subset selected by the IG-GPSO is 98.50%, which is significantly better than the traditional feature selection algorithm. Compared with KNN, the classification effect of the feature subset selected by the IG-GPSO is still optimal. In addition, the results of multiple comparison tests show that the feature selection effect of the IG-GPSO is significantly better than that of traditional feature selection algorithms.Conclusion The feature subset selected by IG-GPSO not only has the best classification effect, but also has the least feature scale (FS). More importantly, the IG-GPSO significantly improves the ACC of SVM in cancer diagnostic.
引用
收藏
页数:17
相关论文
共 50 条
  • [21] Improved salp swarm algorithm based on particle swarm optimization for feature selection
    Ibrahim, Rehab Ali
    Ewees, Ahmed A.
    Oliva, Diego
    Abd Elaziz, Mohamed
    Lu, Songfeng
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2019, 10 (08) : 3155 - 3169
  • [22] Improved salp swarm algorithm based on particle swarm optimization for feature selection
    Rehab Ali Ibrahim
    Ahmed A. Ewees
    Diego Oliva
    Mohamed Abd Elaziz
    Songfeng Lu
    Journal of Ambient Intelligence and Humanized Computing, 2019, 10 : 3155 - 3169
  • [23] A Novel Feature Selection Algorithm using Particle Swarm Optimization for Cancer Microarray Data
    Sahu, Barnali
    Mishra, Debahuti
    INTERNATIONAL CONFERENCE ON MODELLING OPTIMIZATION AND COMPUTING, 2012, 38 : 27 - 31
  • [24] A maximum relevance minimum redundancy hybrid feature selection algorithm based on particle swarm optimization
    Yao, Xu
    Wang, Xiao-Dan
    Zhang, Yu-Xi
    Quan, Wen
    Kongzhi yu Juece/Control and Decision, 2013, 28 (03): : 413 - 417
  • [25] HYBRID BINARY DRAGONFLY ENHANCED PARTICLE SWARM OPTIMIZATION ALGORITHM FOR SOLVING FEATURE SELECTION PROBLEMS
    Tawhid, Mohamed A.
    Dsouza, Kevin B.
    MATHEMATICAL FOUNDATIONS OF COMPUTING, 2018, 1 (02): : 181 - 200
  • [26] Community-Grouping Based Particle Swarm. Optimisation Algorithm for Feature Selection
    Qiu, Jianfeng
    Wan, Jiangchuan
    Zhang, Lei
    Cheng, Fan
    Luo, Yongkang
    2020 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2020,
  • [27] Hybrid firefly particle swarm optimisation algorithm for feature selection problems
    Ragab, Mahmoud
    EXPERT SYSTEMS, 2024, 41 (07)
  • [28] Binary Particle Swarm Optimization based Algorithm for Feature Subset Selection
    Chakraborty, Basabi
    ICAPR 2009: SEVENTH INTERNATIONAL CONFERENCE ON ADVANCES IN PATTERN RECOGNITION, PROCEEDINGS, 2009, : 145 - 148
  • [29] Feature selection algorithm based on bare bones particle swarm optimization
    Zhang, Yong
    Gong, Dunwei
    Hu, Ying
    Zhang, Wanqiu
    NEUROCOMPUTING, 2015, 148 : 150 - 157
  • [30] The feature selection method for SVM with discrete particle swarm optimization algorithm
    Peng Xiyuan
    Wu Hongxing
    Peng Yu
    ISTM/2007: 7TH INTERNATIONAL SYMPOSIUM ON TEST AND MEASUREMENT, VOLS 1-7, CONFERENCE PROCEEDINGS, 2007, : 2523 - 2526