Random K-nearest neighbor algorithm with learning process

被引:0
|
作者
Fu Z.-L. [1 ]
Chen X.-Q. [1 ]
Ren W. [1 ]
Yao Y. [1 ]
机构
[1] Chengdu Institute of Computer Applications, University of Chinese Academy of Sciences, Chengdu
关键词
AdaBoost; Bagging ensemble learning; K-nearest neighbor algorithm; machine learning; random K-Nearest neighbor;
D O I
10.13229/j.cnki.jdxbgxb.20220202
中图分类号
学科分类号
摘要
The traditional KNN(K-nearest neighbor)algorithm is a classic machine learning algorithm. This algorithm has no learning process and needs to traverse all the learning samples when classifying,and is time-sensitive and sensitive to the k value. This paper proposes two random KNN algorithms(RKNN) with a learning process,including the SRKNN algorithm on sample Bootstrap sampling and the ARKNN algorithm on sample feature Bootstrap sampling,both of which belong to Bagging ensemble learning. After learning multiple simple KNNs,the voting output results. The algorithm combines the features of the samples to obtain the combined features,and the simple KNN is obtained based on the combined features. It focuses on how to select the optimal combination coefficient of features,and obtains the selection rules and formulas of the optimal combination features for the best classification accuracy. The RKNN algorithm introduces learning when constructing a simple KNN,it no longer needs to traverse all the learning samples when classifying,but only needs to use the binary search method,and its classification time complexity is an order of magnitude lower than that of the traditional KNN algorithm. The classification accuracy of the RKNN algorithm is significantly improved than that of the traditional KNN algorithm. The RKNN algorithm solves the problem that it is difficult to select the k value using the KNN algorithm. Both theoretical analysis and experimental results show that the proposed RKNN algorithm is an efficient improvement to the KNN algorithm. © 2024 Editorial Board of Jilin University. All rights reserved.
引用
收藏
页码:209 / 220
页数:11
相关论文
共 27 条
  • [1] Cover T M, Hart P E., Nearst neighbor pattern classification, IEEE Transactions on Information Theory, 13, 1, pp. 21-27, (1967)
  • [2] Hart P E., The condensed nearest neighbor rule[J], IEEE Trans actions on Information Theory, 14, 3, pp. 515-516, (1968)
  • [3] Li Rong-lu, Hu Yun-Fa, A density-based method for reducing the amount of training data in kNN text classification, Journal of Computer Research and Development, 41, 4, pp. 539-545, (2004)
  • [4] Zhang Xiao-fei, Huang He-yan, An improved KNN text categorization algorithm by adopting cluster technology, Pattern Recognition and Artificial Intelligence, 22, 6, pp. 936-940, (2009)
  • [5] Hwang W J, Wen K W., Fast KNN classification algorithm based on partial distance search, Electron Letter, 34, 21, pp. 2062-2063, (1998)
  • [6] Pan J S, Qiao Y L, Sun S H., A fast k-nearest neighbors classification algorithm, IEICE Transactions Fundamentals, 87, 4, pp. 961-961, (2004)
  • [7] Samet H., k-Nearst neighbor finding using MaxNearst-Dist, IEEE Transactions on Pattern Analysis and Machine Intelligence, 30, 2, pp. 243-252, (2008)
  • [8] Guo G D, Wang H, Bell D, Et al., Using KNN model for automatic text categorization, Soft Computing-A Fusion of Foundations, Methodologies and Application, 10, 5, pp. 423-430, (2006)
  • [9] Zhu Fu-bao, Xie Li-jie, Tang meng-meng, Et al., Improved KNN classification algorithm based fuzzy C-means, Journal of Central China Normal University (Natural Science Edition), 51, 6, pp. 754-759, (2017)
  • [10] Samet H., The Design an Analysis of Spatial Data Structure, (1990)