Random kernel k-nearest neighbors regression

被引:4
|
作者
Srisuradetchai, Patchanok [1 ]
Suksrikran, Korn [1 ]
机构
[1] Thammasat Univ, Dept Math & Stat, Pathum Thani, Thailand
来源
FRONTIERS IN BIG DATA | 2024年 / 7卷
关键词
bootstrapping; feature selection; k-nearest neighbors regression; kernel k-nearest neighbors; state-of-the-art (SOTA); MOLECULAR DESCRIPTORS; NEURAL-NETWORKS; ENSEMBLE; ALGORITHM; MODEL; SET;
D O I
10.3389/fdata.2024.1402384
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The k-nearest neighbors (KNN) regression method, known for its nonparametric nature, is highly valued for its simplicity and its effectiveness in handling complex structured data, particularly in big data contexts. However, this method is susceptible to overfitting and fit discontinuity, which present significant challenges. This paper introduces the random kernel k-nearest neighbors (RK-KNN) regression as a novel approach that is well-suited for big data applications. It integrates kernel smoothing with bootstrap sampling to enhance prediction accuracy and the robustness of the model. This method aggregates multiple predictions using random sampling from the training dataset and selects subsets of input variables for kernel KNN (K-KNN). A comprehensive evaluation of RK-KNN on 15 diverse datasets, employing various kernel functions including Gaussian and Epanechnikov, demonstrates its superior performance. When compared to standard KNN and the random KNN (R-KNN) models, it significantly reduces the root mean square error (RMSE) and mean absolute error, as well as improving R-squared values. The RK-KNN variant that employs a specific kernel function yielding the lowest RMSE will be benchmarked against state-of-the-art methods, including support vector regression, artificial neural networks, and random forests.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] AutoML for Stream k-Nearest Neighbors Classification
    Bahri, Maroua
    Veloso, Bruno
    Bifet, Albert
    Gama, Joao
    2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 597 - 602
  • [32] Maximizing Reverse k-Nearest Neighbors for Trajectories
    Al Rahat, Tamjid
    Arman, Arif
    Ali, Mohammed Eunus
    DATABASES THEORY AND APPLICATIONS, ADC 2018, 2018, 10837 : 262 - 274
  • [33] The research on an adaptive k-nearest neighbors classifier
    Yu, Xiaopeng
    Yu, Xiaogao
    PROCEEDINGS OF THE FIFTH IEEE INTERNATIONAL CONFERENCE ON COGNITIVE INFORMATICS, VOLS 1 AND 2, 2006, : 535 - 540
  • [34] Heuristics for Computing k-Nearest Neighbors Graphs
    Chavez, Edgar
    Luduena, Veronica
    Reyes, Nora
    COMPUTER SCIENCE - CACIC 2019, 2020, 1184 : 234 - 249
  • [35] Forecasting Earnings Using k-Nearest Neighbors
    Easton, Peter D.
    Kapons, Martin M.
    Monahan, Steven J.
    Schutt, Harm H.
    Weisbrod, Eric H.
    ACCOUNTING REVIEW, 2024, 99 (03): : 115 - 140
  • [36] Ensembles of K-Nearest Neighbors and Dimensionality Reduction
    Okun, Oleg
    Priisalu, Helen
    2008 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-8, 2008, : 2032 - +
  • [37] An Interval Valued K-Nearest Neighbors Classifier
    Derrac, Joaquin
    Chiclana, Francisco
    Garcia, Salvador
    Herrera, Francisco
    PROCEEDINGS OF THE 2015 CONFERENCE OF THE INTERNATIONAL FUZZY SYSTEMS ASSOCIATION AND THE EUROPEAN SOCIETY FOR FUZZY LOGIC AND TECHNOLOGY, 2015, 89 : 378 - 384
  • [38] Hypersphere anchor loss for K-Nearest neighbors
    Ye, Xiang
    He, Zihang
    Wang, Heng
    Li, Yong
    APPLIED INTELLIGENCE, 2023, 53 (24) : 30319 - 30328
  • [39] ANALYSIS OF CUSTOMER CHURN PREDICTION USING LOGISTIC REGRESSION, k-NEAREST NEIGHBORS, DECISION TREE AND RANDOM FOREST ALGORITHMS
    Atay, Mehmet Tarik
    Turanli, Munevver
    ADVANCES AND APPLICATIONS IN STATISTICS, 2025, 92 (02) : 147 - 169
  • [40] A New Algorithm for Large-Scale Geographically Weighted Regression with K-Nearest Neighbors
    Yang, Xiaoyue
    Yang, Yi
    Xu, Shenghua
    Han, Jiakuan
    Chai, Zhengyuan
    Yang, Gang
    ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION, 2023, 12 (07)