On semi-supervised learning

被引:2
|
作者
Cholaquidis, A. [1 ]
Fraiman, R. [1 ]
Sued, M. [2 ]
机构
[1] Univ Republica, Fac Ciencias, Montevideo, Uruguay
[2] INst Calculo, Fac Ciencias Exactas & Nat, Buenos Aires, DF, Argentina
关键词
Semi-supervised learning; Small training sample; Consistency; PATTERN-RECOGNITION; ERROR;
D O I
10.1007/s11749-019-00690-2
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Major efforts have been made, mostly in the machine learning literature, to construct good predictors combining unlabelled and labelled data. These methods are known as semi-supervised. They deal with the problem of how to take advantage, if possible, of a huge amount of unlabelled data to perform classification in situations where there are few labelled data. This is not always feasible: it depends on the possibility to infer the labels from the unlabelled data distribution. Nevertheless, several algorithms have been proposed recently. In this work, we present a new method that, under almost necessary conditions, attains asymptotically the performance of the best theoretical rule when the size of the unlabelled sample goes to infinity, even if the size of the labelled sample remains fixed. Its performance and computational time are assessed through simulations and in the well- known "Isolet" real data of phonemes, where a strong dependence on the choice of the initial training sample is shown. The main focus of this work is to elucidate when and why semi-supervised learning works in the asymptotic regime described above. The set of necessary assumptions, although reasonable, show that semi-parametric methods only attain consistency for very well-conditioned problems.
引用
收藏
页码:914 / 937
页数:24
相关论文
共 50 条
  • [21] Universal Semi-Supervised Learning
    Huang, Zhuo
    Xue, Chao
    Han, Bo
    Yang, Jian
    Gong, Chen
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [22] Adversarial Dropout for Supervised and Semi-Supervised Learning
    Park, Sungrae
    Park, JunKeon
    Shin, Su-Jin
    Moon, Il-Chul
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 3917 - 3924
  • [23] Supervised and semi-supervised machine learning ranking
    Vittaut, Jean-Noel
    Gallinari, Patrick
    COMPARATIVE EVALUATION OF XML INFORMATION RETRIEVAL SYSTEMS, 2007, 4518 : 213 - 222
  • [24] Semi-Supervised Learning via Regularized Boosting Working on Multiple Semi-Supervised Assumptions
    Chen, Ke
    Wang, Shihai
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2011, 33 (01) : 129 - 143
  • [25] Semi-supervised Neighborhood Preserving Discriminant Embedding: A Semi-supervised Subspace Learning Algorithm
    Mehdizadeh, Maryam
    MacNish, Cara
    Khan, R. Nazim
    Bennamoun, Mohammed
    COMPUTER VISION - ACCV 2010, PT III, 2011, 6494 : 199 - +
  • [26] Semi-supervised metric learning via topology preserving multiple semi-supervised assumptions
    Wang, Qianying
    Yuen, Pong C.
    Feng, Guocan
    PATTERN RECOGNITION, 2013, 46 (09) : 2576 - 2587
  • [27] Efficiently Learning the Graph for Semi-supervised Learning
    Sharma, Dravyansh
    Jones, Maxwell
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 1900 - 1910
  • [28] Adaptive Active Learning for Semi-supervised Learning
    Li Y.-C.
    Xiao F.
    Chen Z.
    Li B.
    Ruan Jian Xue Bao/Journal of Software, 2020, 31 (12): : 3808 - 3822
  • [29] POSITIVE UNLABELED LEARNING BY SEMI-SUPERVISED LEARNING
    Wang, Zhuowei
    Jiang, Jing
    Long, Guodong
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 2976 - 2980
  • [30] Broad learning system for semi-supervised learning
    Liu, Zheng
    Huang, Shiluo
    Jin, Wei
    Mu, Ying
    NEUROCOMPUTING, 2021, 444 (444) : 38 - 47