kNNVWC: An Efficient k-Nearest Neighbors Approach Based on Various-Widths Clustering

被引:31
|
作者
Almalawi, Abdul Mohsen [1 ]
Fahad, Adil [2 ]
Tari, Zahir [3 ]
Cheema, Muhammad Aamir [4 ]
Khalil, Ibrahim [3 ]
机构
[1] King Abdulaziz Univ, Sch Comp Sci & Informat Technol, Jeddah 21413, Saudi Arabia
[2] Al Baha Univ, Coll Comp Sci & Informat Technol, Dept Comp Sci, Al Baha, Saudi Arabia
[3] RMIT Univ, Sch Comp Sci & Informat Technol CSIT, Distributed Syst & Networking DSN Discipline, Melbourne, Vic 3000, Australia
[4] Monash Univ, Fac Informat Technol, Clayton, Vic 3168, Australia
基金
澳大利亚研究理事会;
关键词
Clustering; K-nearest neighbour; high dimensionality; performance; SCADA; DISTANCE-BASED OUTLIERS; ALGORITHM; SEARCH;
D O I
10.1109/TKDE.2015.2460735
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The k-nearest neighbor approach (k-NN) has been extensively used as a powerful non-parametric technique in many scientific and engineering applications. However, this approach incurs a large computational cost. Hence, this issue has become an active research field. In this work, a novel k-NN approach based on various-widths clustering, named kNNVWC, to efficiently find k-NNs for a query object from a given data set, is presented. kNNVWC does clustering using various widths, where a data set is clustered with a global width first and each produced cluster that meets the predefined criteria is recursively clustered with its own local width that suits its distribution. This reduces the clustering time, in addition to balancing the number of produced clusters and their respective sizes. Maximum efficiency is achieved by using triangle inequality to prune unlikely clusters. Experimental results demonstrate that kNNVWC performs well in finding k-NNs for query objects compared to a number of k-NN search algorithms, especially for a data set with high dimensions, various distributions and large size.
引用
收藏
页码:68 / 81
页数:14
相关论文
共 50 条
  • [21] Density Peaks Clustering Algorithm Based on Representative Points and K-nearest Neighbors
    Zhang Q.-H.
    Zhou J.-P.
    Dai Y.-Y.
    Wang G.-Y.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (12): : 5629 - 5648
  • [22] Adaptive density peak clustering based on K-nearest neighbors with aggregating strategy
    Liu Yaohui
    Ma Zhengming
    Yu Fang
    KNOWLEDGE-BASED SYSTEMS, 2017, 133 : 208 - 220
  • [23] Density peaks clustering based on k-nearest neighbors and self-recommendation
    Sun, Lin
    Qin, Xiaoying
    Ding, Weiping
    Xu, Jiucheng
    Zhang, Shiguang
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2021, 12 (07) : 1913 - 1938
  • [24] Local density based on weighted K-nearest neighbors for density peaks clustering
    Ding, Sifan
    Li, Min
    Huang, Tianyi
    Zhu, William
    KNOWLEDGE-BASED SYSTEMS, 2024, 305
  • [25] Density peaks clustering based on k-nearest neighbors and self-recommendation
    Lin Sun
    Xiaoying Qin
    Weiping Ding
    Jiucheng Xu
    Shiguang Zhang
    International Journal of Machine Learning and Cybernetics, 2021, 12 : 1913 - 1938
  • [26] Fast agglomerative clustering using information of k-nearest neighbors
    Chang, Chih-Tang
    Lai, Jim Z. C.
    Jeng, M. D.
    PATTERN RECOGNITION, 2010, 43 (12) : 3958 - 3968
  • [27] K-Nearest Neighbors Hashing
    He, Xiangyu
    Wang, Peisong
    Cheng, Jian
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2834 - 2843
  • [28] Modernizing k-nearest neighbors
    Elizabeth Yancey, Robin
    Xin, Bochao
    Matloff, Norm
    STAT, 2021, 10 (01):
  • [29] A new approach for increasing K-nearest neighbors performance
    Aamer, Youssef
    Benkaouz, Yahya
    Ouzzif, Mohammed
    Bouragba, Khalid
    2020 8TH INTERNATIONAL CONFERENCE ON WIRELESS NETWORKS AND MOBILE COMMUNICATIONS (WINCOM 2020), 2020, : 35 - 39
  • [30] A k-nearest neighbors approach to the design of radar detectors
    Coluccia, Angelo
    Fascista, Alessio
    Ricci, Giuseppe
    SIGNAL PROCESSING, 2020, 174 (174)