WHHO: enhanced Harris hawks optimizer for feature selection in high-dimensional data

被引:1
|
作者
Meilin Zhang [1 ]
Huiling Chen [1 ]
Ali Asghar Heidari [2 ]
Yi Chen [1 ]
Zongda Wu [4 ]
Zhennao Cai [1 ]
Lei Liu [3 ]
机构
[1] Wenzhou University,Department of Computer Science and Artificial Intelligence
[2] University of Tehran,School of Surveying and Geospatial Engineering, College of Engineering
[3] Sichuan University,College of Computer Science
[4] Shaoxing University,Department of Computer Science and Engineering
关键词
Harris hawks optimization; Wormhole strategy; Swarm intelligence; High dimensional feature selection;
D O I
10.1007/s10586-024-04770-3
中图分类号
学科分类号
摘要
The Harris Hawks optimization (HHO) algorithm, categorized as a powerful meta-heuristic algorithm (MA), has gained extensive usage because it lacks parameter adjustment and shows powerful optimization ability. At the same time, it cannot be ignored within the realm of feature selection. However, HHO is deficient in its capacity to escape local optima, which deserves attention. HHO did not perform enough local search in the exploitation stage, resulting in poor final results. This limitation occurs from time to time when facing the high-dimensional feature selection. To address these concerns, a modification of HHO incorporating the wormhole strategy (WS) is proposed, called WHHO. As an example of local search, WS can mutate around the current and optimal search positions. By ensuring population diversity and enhancing the algorithm's capability to escape from local optima, WS contributes to improving results accuracy. WHHO first conducted qualitative experiments at CEC 2017, including historical trajectory analysis, balance analysis, and diversity analysis. Then, WHHO was compared with seven classical MAs, seven advanced MAs, and five well-known HHO variants. In the three groups of experiments, careful consideration of mean, standard deviation, Friedman test, and Wilcoxon sign-rank test results reveals that WHHO secures the top ranking. Furthermore, the binary variant of the algorithm, BWHHO, combined with the k nearest neighbor classifier, was developed for high-dimensional feature selection. Compared with nine binary MAs, WHHO uses fewer features to predict more accurate results on twelve high-dimensional datasets. The fitness value of WHHO on twelve datasets won the overwhelming first place, surpassing the second place with an average ranking of 2.5833. WHHO's ability to reduce the number of selected features is particularly impressive while maintaining high prediction accuracy. In the face of thousands of features, WHHO consistently selects fewer features, with retention rates below 20% on all datasets. This highlights WHHO's optimized feature selection ability and its effectiveness in simplifying the prediction process. Therefore, WHHO is an effective improved optimizer, especially for dealing with high-dimensional feature selection, which has effective value.
引用
收藏
相关论文
共 50 条
  • [21] A filter feature selection for high-dimensional data
    Janane, Fatima Zahra
    Ouaderhman, Tayeb
    Chamlal, Hasna
    JOURNAL OF ALGORITHMS & COMPUTATIONAL TECHNOLOGY, 2023, 17
  • [22] A Clustering-Guided Integer Brain Storm Optimizer for Feature Selection in High-Dimensional Data
    Jia Yun-Tao
    Zhang Wan-Qiu
    He Chun-Lin
    DISCRETE DYNAMICS IN NATURE AND SOCIETY, 2021, 2021
  • [23] Feature selection for high-dimensional temporal data
    Michail Tsagris
    Vincenzo Lagani
    Ioannis Tsamardinos
    BMC Bioinformatics, 19
  • [24] Feature Selection with High-Dimensional Imbalanced Data
    Van Hulse, Jason
    Khoshgoftaar, Taghi M.
    Napolitano, Amri
    Wald, Randall
    2009 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW 2009), 2009, : 507 - 514
  • [25] Feature selection for high-dimensional temporal data
    Tsagris, Michail
    Lagani, Vincenzo
    Tsamardinos, Ioannis
    BMC BIOINFORMATICS, 2018, 19
  • [26] FEATURE SELECTION FOR HIGH-DIMENSIONAL DATA ANALYSIS
    Verleysen, Michel
    ECTA 2011/FCTA 2011: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON EVOLUTIONARY COMPUTATION THEORY AND APPLICATIONS AND INTERNATIONAL CONFERENCE ON FUZZY COMPUTATION THEORY AND APPLICATIONS, 2011,
  • [27] Enhanced Harris Hawks optimization as a feature selection for the prediction of student performance
    Hamza Turabieh
    Sana Al Azwari
    Mahmoud Rokaya
    Wael Alosaimi
    Abdullah Alharbi
    Wajdi Alhakami
    Mrim Alnfiai
    Computing, 2021, 103 : 1417 - 1438
  • [28] Enhanced Harris Hawks optimization as a feature selection for the prediction of student performance
    Turabieh, Hamza
    Al Azwari, Sana
    Rokaya, Mahmoud
    Alosaimi, Wael
    Alharbi, Abdullah
    Alhakami, Wajdi
    Alnfiai, Mrim
    COMPUTING, 2021, 103 (07) : 1417 - 1438
  • [29] Binary dwarf mongoose optimizer for solving high-dimensional feature selection problems
    Akinola, Olatunji A.
    Agushaka, Jeffrey O.
    Ezugwu, Absalom E.
    PLOS ONE, 2022, 17 (10):
  • [30] Improving feature selection performance for classification of gene expression data using Harris Hawks optimizer with variable neighborhood learning
    Qu, Chiwen
    Zhang, Lupeng
    Li, Jinlong
    Deng, Fang
    Tang, Yifan
    Zeng, Xiaomin
    Peng, Xiaoning
    BRIEFINGS IN BIOINFORMATICS, 2021, 22 (05)