WHHO: enhanced Harris hawks optimizer for feature selection in high-dimensional data

被引:1
|
作者
Meilin Zhang [1 ]
Huiling Chen [1 ]
Ali Asghar Heidari [2 ]
Yi Chen [1 ]
Zongda Wu [4 ]
Zhennao Cai [1 ]
Lei Liu [3 ]
机构
[1] Wenzhou University,Department of Computer Science and Artificial Intelligence
[2] University of Tehran,School of Surveying and Geospatial Engineering, College of Engineering
[3] Sichuan University,College of Computer Science
[4] Shaoxing University,Department of Computer Science and Engineering
关键词
Harris hawks optimization; Wormhole strategy; Swarm intelligence; High dimensional feature selection;
D O I
10.1007/s10586-024-04770-3
中图分类号
学科分类号
摘要
The Harris Hawks optimization (HHO) algorithm, categorized as a powerful meta-heuristic algorithm (MA), has gained extensive usage because it lacks parameter adjustment and shows powerful optimization ability. At the same time, it cannot be ignored within the realm of feature selection. However, HHO is deficient in its capacity to escape local optima, which deserves attention. HHO did not perform enough local search in the exploitation stage, resulting in poor final results. This limitation occurs from time to time when facing the high-dimensional feature selection. To address these concerns, a modification of HHO incorporating the wormhole strategy (WS) is proposed, called WHHO. As an example of local search, WS can mutate around the current and optimal search positions. By ensuring population diversity and enhancing the algorithm's capability to escape from local optima, WS contributes to improving results accuracy. WHHO first conducted qualitative experiments at CEC 2017, including historical trajectory analysis, balance analysis, and diversity analysis. Then, WHHO was compared with seven classical MAs, seven advanced MAs, and five well-known HHO variants. In the three groups of experiments, careful consideration of mean, standard deviation, Friedman test, and Wilcoxon sign-rank test results reveals that WHHO secures the top ranking. Furthermore, the binary variant of the algorithm, BWHHO, combined with the k nearest neighbor classifier, was developed for high-dimensional feature selection. Compared with nine binary MAs, WHHO uses fewer features to predict more accurate results on twelve high-dimensional datasets. The fitness value of WHHO on twelve datasets won the overwhelming first place, surpassing the second place with an average ranking of 2.5833. WHHO's ability to reduce the number of selected features is particularly impressive while maintaining high prediction accuracy. In the face of thousands of features, WHHO consistently selects fewer features, with retention rates below 20% on all datasets. This highlights WHHO's optimized feature selection ability and its effectiveness in simplifying the prediction process. Therefore, WHHO is an effective improved optimizer, especially for dealing with high-dimensional feature selection, which has effective value.
引用
收藏
相关论文
共 50 条
  • [41] On the scalability of feature selection methods on high-dimensional data
    Bolon-Canedo, V.
    Rego-Fernandez, D.
    Peteiro-Barral, D.
    Alonso-Betanzos, A.
    Guijarro-Berdinas, B.
    Sanchez-Marono, N.
    KNOWLEDGE AND INFORMATION SYSTEMS, 2018, 56 (02) : 395 - 442
  • [42] A hybrid feature selection scheme for high-dimensional data
    Ganjei, Mohammad Ahmadi
    Boostani, Reza
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 113
  • [43] Evaluating Feature Selection Robustness on High-Dimensional Data
    Pes, Barbara
    HYBRID ARTIFICIAL INTELLIGENT SYSTEMS (HAIS 2018), 2018, 10870 : 235 - 247
  • [44] Feature selection for classifying high-dimensional numerical data
    Wu, YM
    Zhang, AD
    PROCEEDINGS OF THE 2004 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 2, 2004, : 251 - 258
  • [45] An Efficient High-dimensional Feature Selection Approach Driven By Enhanced Multi-strategy Grey Wolf Optimizer for Biological Data Classification
    Mafarja, Majdi
    Thaher, Thaer
    Too, Jingwei
    Chantar, Hamouda
    Turabieh, Hamza
    Houssein, Essam H.
    Emam, Marwa M.
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (02): : 1749 - 1775
  • [46] An Efficient High-dimensional Feature Selection Approach Driven By Enhanced Multi-strategy Grey Wolf Optimizer for Biological Data Classification
    Majdi Mafarja
    Thaer Thaher
    Jingwei Too
    Hamouda Chantar
    Hamza Turabieh
    Essam H. Houssein
    Marwa M. Emam
    Neural Computing and Applications, 2023, 35 : 1749 - 1775
  • [47] A Particle Swarm Optimizer with Biased Exploration and Exploitation for High-Dimensional Optimization and Feature Selection
    Chen, Minchong
    Li, Hong
    Tu, Jiwei
    He, Bin
    Yin, Zhen
    Hou, Xuejing
    Yu, Qi
    UNMANNED SYSTEMS, 2024,
  • [48] Lens-imaging learning Harris hawks optimizer for global optimization and its application to feature selection
    Long, Wen
    Jiao, Jianjun
    Xu, Ming
    Tang, Mingzhu
    Wu, Tiebin
    Cai, Shaohong
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 202
  • [49] A Light Causal Feature Selection Approach to High-Dimensional Data
    Ling, Zhaolong
    Li, Ying
    Zhang, Yiwen
    Yu, Kui
    Zhou, Peng
    Li, Bo
    Wu, Xindong
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (08) : 7639 - 7650
  • [50] Filter Feature Selection Performance Comparison in High-dimensional Data
    Huertas, Carlos
    Juarez-Ramirez, Reyes
    2014 17TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION), 2014,