On semi-supervised learning

被引:2
|
作者
Cholaquidis, A. [1 ]
Fraiman, R. [1 ]
Sued, M. [2 ]
机构
[1] Univ Republica, Fac Ciencias, Montevideo, Uruguay
[2] INst Calculo, Fac Ciencias Exactas & Nat, Buenos Aires, DF, Argentina
关键词
Semi-supervised learning; Small training sample; Consistency; PATTERN-RECOGNITION; ERROR;
D O I
10.1007/s11749-019-00690-2
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Major efforts have been made, mostly in the machine learning literature, to construct good predictors combining unlabelled and labelled data. These methods are known as semi-supervised. They deal with the problem of how to take advantage, if possible, of a huge amount of unlabelled data to perform classification in situations where there are few labelled data. This is not always feasible: it depends on the possibility to infer the labels from the unlabelled data distribution. Nevertheless, several algorithms have been proposed recently. In this work, we present a new method that, under almost necessary conditions, attains asymptotically the performance of the best theoretical rule when the size of the unlabelled sample goes to infinity, even if the size of the labelled sample remains fixed. Its performance and computational time are assessed through simulations and in the well- known "Isolet" real data of phonemes, where a strong dependence on the choice of the initial training sample is shown. The main focus of this work is to elucidate when and why semi-supervised learning works in the asymptotic regime described above. The set of necessary assumptions, although reasonable, show that semi-parametric methods only attain consistency for very well-conditioned problems.
引用
收藏
页码:914 / 937
页数:24
相关论文
共 50 条
  • [41] Augmentation Learning for Semi-Supervised Classification
    Frommknecht, Tim
    Zipf, Pedro Alves
    Fan, Quanfu
    Shvetsova, Nina
    Kuehne, Hilde
    PATTERN RECOGNITION, DAGM GCPR 2022, 2022, 13485 : 85 - 98
  • [42] Semi-Supervised Learning with Normalizing Flows
    Izmailov, Pavel
    Kirichenko, Polina
    Finzi, Marc
    Wilson, Andrew Gordon
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [43] Semi-Supervised Learning with Scarce Annotations
    Rebuffi, Sylvestre-Alvise
    Ehrhardt, Sebastien
    Han, Kai
    Vedaldi, Andrea
    Zisserman, Andrew
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 3294 - 3302
  • [44] Unbiased generative semi-supervised learning
    1600, Microtome Publishing (15):
  • [45] Wasserstein Propagation for Semi-Supervised Learning
    Solomon, Justin
    Rustamov, Raif M.
    Guibas, Leonidas
    Butscher, Adrian
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 1), 2014, 32
  • [46] Unbiased Generative Semi-Supervised Learning
    Fox-Roberts, Patrick
    Rosten, Edward
    JOURNAL OF MACHINE LEARNING RESEARCH, 2014, 15 : 367 - 443
  • [47] Normalized LDA for Semi-supervised Learning
    Fan, Bin
    Lei, Zhen
    Li, Stan Z.
    2008 8TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE & GESTURE RECOGNITION (FG 2008), VOLS 1 AND 2, 2008, : 416 - +
  • [48] FMixCutMatch for semi-supervised deep learning
    Wei, Xiang
    Wei, Xiaotao
    Kong, Xiangyuan
    Lu, Siyang
    Xing, Weiwei
    Lu, Wei
    Neural Networks, 2021, 133 : 166 - 176
  • [49] Quantum annealing for semi-supervised learning
    郑玉鳞
    张文
    周诚
    耿巍
    Chinese Physics B, 2021, (04) : 91 - 97
  • [50] Semi-Supervised Learning on Riemannian Manifolds
    Mikhail Belkin
    Partha Niyogi
    Machine Learning, 2004, 56 : 209 - 239