Self-training: A survey

被引:1
|
作者
Amini, Massih-Reza [1 ]
Feofanov, Vasilii [1 ]
Pauletto, Loic [1 ]
Hadjadj, Lies [1 ]
Devijver, Emilie [1 ]
Maximov, Yury [2 ]
机构
[1] Univ Grenoble Alpes, CNRS, Lab Informat Grenoble, Grenoble, France
[2] Los Alamos Natl Lab, Theoret Div, Los Alamos, NM USA
关键词
Semi-supervised learning; Self-training;
D O I
10.1016/j.neucom.2024.128904
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-training methods have gained significant attention in recent years due to their effectiveness in leveraging small labeled datasets and large unlabeled observations for prediction tasks. These models identify decision boundaries in low-density regions without additional assumptions about data distribution, using the confidence scores of a learned classifier. The core principle of self-training involves iteratively assigning pseudo-labels to unlabeled samples with confidence scores above a certain threshold, enriching the labeled dataset and retraining the classifier. This paper presents self-training methods for binary and multi-class classification, along with variants and related approaches such as consistency-based methods and transductive learning. We also briefly describe self-supervised learning and reinforced self-training. Furthermore, we highlight popular applications of self-training and discuss the importance of dynamic thresholding and reducing pseudo-label noise for performance improvement. To the best of our knowledge, this is the first thorough and complete survey on self-training.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Online Continual Adaptation with Active Self-Training
    Zhou, Shiji
    Zhao, Han
    Zhang, Shanghang
    Wang, Lianzhe
    Chang, Heng
    Wang, Zhi
    Zhu, Wenwu
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [42] A self-training spiking superconducting neuromorphic architecture
    M. L. Schneider
    E. M. Jué
    M. R. Pufall
    K. Segall
    C. W. Anderson
    npj Unconventional Computing, 2 (1):
  • [43] Saliency Regularization for Self-Training with Partial Annotations
    Wang, Shouwen
    Wan, Qian
    Xiang, Xiang
    Zeng, Zhigang
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 1611 - 1620
  • [44] A Soft-Labeled Self-Training Approach
    Mey, Alexander
    Loog, Marco
    2016 23RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2016, : 2604 - 2609
  • [45] Self-training for handwritten word recognition and retrieval
    Wolf, Fabian
    Fink, Gernot A.
    INTERNATIONAL JOURNAL ON DOCUMENT ANALYSIS AND RECOGNITION, 2024, 27 (03) : 225 - 244
  • [46] Pedestrian Classification Using Self-Training Algorithm
    Jiralerspong, Trongmun
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 515 - 520
  • [47] Double Pressure Presentation for Calligraphy Self-training
    Morikawa, Ami
    Tsuda, Naoaki
    Nomura, Yoshihiko
    Kato, Norihiko
    COMPANION OF THE 2018 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION (HRI'18), 2018, : 199 - 200
  • [48] SELF-TRAINING AND PRE-TRAINING ARE COMPLEMENTARY FOR SPEECH RECOGNITION
    Xu, Qiantong
    Baevski, Alexei
    Likhomanenko, Tatiana
    Tomasello, Paden
    Conneau, Alexis
    Collobert, Ronan
    Synnaeve, Gabriel
    Auli, Michael
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3030 - 3034
  • [49] Improving Skin Lesion Segmentation with Self-Training
    Dzieniszewska, Aleksandra
    Garbat, Piotr
    Piramidowicz, Ryszard
    CANCERS, 2024, 16 (06)
  • [50] A Self-Training Approach for Short Text Clustering
    Hadifar, Amir
    Sterckx, Lucas
    Demeester, Thomas
    Develder, Chris
    4TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP (REPL4NLP-2019), 2019, : 194 - 199