Research on Improvements of Feature Selection Using Forest Optimization Algorithm

被引:0
|
作者
Chu B. [1 ,2 ]
Li Z.-S. [1 ,2 ]
Zhang M.-L. [1 ,2 ]
Yu H.-H. [1 ,2 ]
机构
[1] College of Computer Science and Technology, Jilin University, Changchun
[2] Key Laboratory of Symbolic Computation and Knowledge Engineering, Jilin University, Ministry of Education (Jilin University), Changchun
来源
Ruan Jian Xue Bao/Journal of Software | 2018年 / 29卷 / 09期
基金
中国国家自然科学基金;
关键词
Feature selection; Greedy strategy; IFSFOA; Initialization; Updating mechanism;
D O I
10.13328/j.cnki.jos.005395
中图分类号
学科分类号
摘要
In classification, feature selection has been an important, but difficult problem. Recent research results disclosed that feature selection using forest optimization algorithm (FSFOA) has a better classification performance and good dimensionality reduction ability. However, the randomness of initialization phase, the limitations of updating mechanism and the inferior quality of the new tree in the local seeding stage severely limit the classification performance and dimensionality reduction ability of the algorithm. In this paper, a new initialization strategy and updating mechanism are used and a greedy strategy is added in the local seeding stage to form a new feature selection algorithm (IFSFOA) in order to maximize the classification performance and simultaneously minimize the number of features. In experiment, IFSFOA uses SVM, J48 and KNN classifiers to guide the learning process while utilizing the machine learning database UCI for testing. The results show that compared with FSFOA, IFSFOA has a significant improvement in classification performance and dimensionality reduction. Comparing IFSFOA algorithm with more efficient feature selection methods proposed in recent years, IFSFOA is still very competitive in both accuracy and dimensionality reduction. © Copyright 2018, Institute of Software, the Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:2547 / 2558
页数:11
相关论文
共 30 条
  • [1] Zhou Z.H., Machine Learning, pp. 247-261, (2016)
  • [2] Guyon I., Elisseeff A., An introduction to variable and feature selection, Journal of Machine Learning Research, 3, 6, pp. 1157-1182, (2003)
  • [3] Ghaemi M., Feizi-Derakhshi M.R., Feature selection using forest optimization algorithm, Pattern Recognition, 60, pp. 121-129, (2016)
  • [4] Xue B., Zhang M., Browne W.N., Novel initialisation and updating mechanisms in PSO for feature selection in classification, Proc. of the European Conf. on Applications of Evolutionary Computation, pp. 428-438, (2013)
  • [5] Nie D.G., Research on improvements and discretization of forest optimization algorithm, (2016)
  • [6] Tabakhi S., Moradi P., Akhlaghian F., An unsupervised feature selection algorithm based on ant colony optimization, Engineering Applications of Artificial Intelligence, 32, 6, pp. 112-123, (2014)
  • [7] Zhu W., Si G., Zhang Y., Et al., Neighborhood effective information ratio for hybrid feature subset evaluation and selection, Neurocomputing, 99, pp. 25-37, (2013)
  • [8] Xue B., Zhang M., Browne W.N., Particle swarm optimisation for feature selection in classification, Applied Soft Computing, 18, C, pp. 261-276, (2014)
  • [9] Hu Q., Che X., Zhang L., Et al., Feature evaluation and selection based on neighborhood soft margin, Neurocomputing, 73, 10-12, pp. 2114-2124, (2010)
  • [10] Moustakidis S.P., Theocharis J.B., SVM-FuzCoC: A novel SVM-based feature selection method using a fuzzy complementary criterion, Pattern Recognition, 43, 11, pp. 3712-3729, (2010)