Improved kNN Rule for Small Training Sets

被引:2
|
作者
Cheamanunkul, Sunsern [1 ]
Freund, Yoav [1 ]
机构
[1] Univ Calif San Diego, Dept Comp Sci & Engn, La Jolla, CA 92093 USA
关键词
D O I
10.1109/ICMLA.2014.37
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The traditional k-NN classification rule predicts a label based on the most common label of the k nearest neighbors (the plurality rule). It is known that the plurality rule is optimal when the number of examples tends to infinity. In this paper we show that the plurality rule is sub-optimal when the number of labels is large and the number of examples is small. We propose a simple k-NN rule that takes into account the labels of all of the neighbors, rather than just the most common label. We present a number of experiments on both synthetic datasets and real-world datasets, including MNIST and SVHN. We show that our new rule can achieve lower error rates compared to the majority rule in many cases.
引用
收藏
页码:201 / 206
页数:6
相关论文
共 50 条
  • [31] Refine and merge: Generating small rule bases from training data
    Sudkamp, T
    Knapp, J
    Knapp, A
    JOINT 9TH IFSA WORLD CONGRESS AND 20TH NAFIPS INTERNATIONAL CONFERENCE, PROCEEDINGS, VOLS. 1-5, 2001, : 197 - 202
  • [32] Handling small training sets confidence/accuracy with regard to new examples
    Caulfield, HJ
    OPTICAL PATTERN RECOGNITION XIV, 2003, 5106 : 194 - 199
  • [33] Domain-independent automatic keyphrase indexing with small training sets
    Medelyan, Ena
    Witten, Ian H.
    JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY, 2008, 59 (07): : 1026 - 1040
  • [34] Neural-network design for small training sets of high dimension
    Yuan, JL
    Fine, TL
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 1998, 9 (02): : 266 - 280
  • [35] Designing compact feedforward neural models with small training data sets
    Greenman, RM
    Stepniewski, SW
    Jorgensen, CC
    Roth, KR
    JOURNAL OF AIRCRAFT, 2002, 39 (03): : 452 - 459
  • [36] Intention Understanding in Small Training Data Sets by Using Transfer Learning
    Joko, Hideaki
    Ucihde, Hayato
    Koji, Yusuke
    Otsuka, Takahiro
    2018 ELEVENTH INTERNATIONAL CONFERENCE ON MOBILE COMPUTING AND UBIQUITOUS NETWORK (ICMU 2018), 2018,
  • [37] Speech Recognition Using Convolutional Neural Networks on Small Training Sets
    Poliyev, A. V.
    Korsun, O. N.
    2019 WORKSHOP ON MATERIALS AND ENGINEERING IN AERONAUTICS, 2020, 714
  • [38] Ensemble based Classification using Small Training sets : A Novel Approach
    Veni, C. V. Krishna
    Rani, T. Sobha
    2014 IEEE SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE IN ENSEMBLE LEARNING (CIEL), 2014, : 13 - 20
  • [39] An ensemble method using small training sets for imbalanced data sets: Application to drugs used for kinases
    Rani, T. Sobha
    Soujanya, P. V.
    2013 SIXTH INTERNATIONAL CONFERENCE ON CONTEMPORARY COMPUTING (IC3), 2013, : 516 - 521
  • [40] An improved compact formulation for the assortment optimization problem with small consideration sets
    Roberti, Roberto
    Salvagnin, Domenico
    Fischetti, Matteo
    JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY, 2025,