Random kernel k-nearest neighbors regression

被引:4
|
作者
Srisuradetchai, Patchanok [1 ]
Suksrikran, Korn [1 ]
机构
[1] Thammasat Univ, Dept Math & Stat, Pathum Thani, Thailand
来源
FRONTIERS IN BIG DATA | 2024年 / 7卷
关键词
bootstrapping; feature selection; k-nearest neighbors regression; kernel k-nearest neighbors; state-of-the-art (SOTA); MOLECULAR DESCRIPTORS; NEURAL-NETWORKS; ENSEMBLE; ALGORITHM; MODEL; SET;
D O I
10.3389/fdata.2024.1402384
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The k-nearest neighbors (KNN) regression method, known for its nonparametric nature, is highly valued for its simplicity and its effectiveness in handling complex structured data, particularly in big data contexts. However, this method is susceptible to overfitting and fit discontinuity, which present significant challenges. This paper introduces the random kernel k-nearest neighbors (RK-KNN) regression as a novel approach that is well-suited for big data applications. It integrates kernel smoothing with bootstrap sampling to enhance prediction accuracy and the robustness of the model. This method aggregates multiple predictions using random sampling from the training dataset and selects subsets of input variables for kernel KNN (K-KNN). A comprehensive evaluation of RK-KNN on 15 diverse datasets, employing various kernel functions including Gaussian and Epanechnikov, demonstrates its superior performance. When compared to standard KNN and the random KNN (R-KNN) models, it significantly reduces the root mean square error (RMSE) and mean absolute error, as well as improving R-squared values. The RK-KNN variant that employs a specific kernel function yielding the lowest RMSE will be benchmarked against state-of-the-art methods, including support vector regression, artificial neural networks, and random forests.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Two steps hybrid calibration algorithm of support vector regression and K-nearest neighbors
    Hamed, Yaman
    Alzahrani, Ahmed Ibrahim
    Shafie, A'fza
    Mustaffa, Zahiraniza
    Ismail, Mokhtar Che
    Eng, Kee Kok
    ALEXANDRIA ENGINEERING JOURNAL, 2020, 59 (03) : 1181 - 1190
  • [42] DETERMINANTS OF CONSUMER EGG PRICES IN TURKIYE USING K-NEAREST NEIGHBORS REGRESSION ANALYSIS
    Cevrimli, M. B.
    Mat, B.
    Polat, M.
    Arikan, M. S.
    Akin, A. Cu.
    Tekindal, M. A.
    Gunlu, A.
    JOURNAL OF ANIMAL AND PLANT SCIENCES-JAPS, 2024, 34 (03): : 557 - 563
  • [43] Distributed architecture for k-nearest neighbors recommender systems
    Formoso, Vreixo
    Fernandez, Diego
    Cacheda, Fidel
    Carneiro, Victor
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2015, 18 (04): : 997 - 1017
  • [44] Compressed kNN: K-Nearest Neighbors with Data Compression
    Salvador-Meneses, Jaime
    Ruiz-Chavez, Zoila
    Garcia-Rodriguez, Jose
    ENTROPY, 2019, 21 (03)
  • [45] Human Sleep Scoring Based on K-Nearest Neighbors
    Qureshi, Shahnawaz
    Karrila, Seppo
    Vanichayobon, Sirirut
    TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2018, 26 (06) : 2802 - +
  • [46] Parallel Search of k-Nearest Neighbors with Synchronous Operations
    Sismanis, Nikos
    Pitsianis, Nikos
    Sun, Xiaobai
    2012 IEEE CONFERENCE ON HIGH PERFORMANCE EXTREME COMPUTING (HPEC), 2012,
  • [47] A hashing strategy for efficient k-nearest neighbors computation
    Vanco, M
    Brunnett, G
    Schreiber, T
    COMPUTER GRAPHICS INTERNATIONAL, PROCEEDINGS, 1999, : 120 - 128
  • [48] EDITING FOR THE K-NEAREST NEIGHBORS RULE BY A GENETIC ALGORITHM
    KUNCHEVA, LI
    PATTERN RECOGNITION LETTERS, 1995, 16 (08) : 809 - 814
  • [49] A new approach for increasing K-nearest neighbors performance
    Aamer, Youssef
    Benkaouz, Yahya
    Ouzzif, Mohammed
    Bouragba, Khalid
    2020 8TH INTERNATIONAL CONFERENCE ON WIRELESS NETWORKS AND MOBILE COMMUNICATIONS (WINCOM 2020), 2020, : 35 - 39
  • [50] k-Nearest Neighbors for automated classification of celestial objects
    LiLi Li
    YanXia Zhang
    YongHeng Zhao
    Science in China Series G: Physics, Mechanics and Astronomy, 2008, 51 : 916 - 922