An improved sample selection framework for learning with noisy labels

被引:0
|
作者
Zhang, Qian [1 ]
Zhu, Yi [1 ]
Yang, Ming [2 ]
Jin, Ge [1 ]
Zhu, Yingwen [1 ]
Lu, Yanjun [1 ]
Zou, Yu [1 ,3 ]
Chen, Qiu [4 ]
机构
[1] Jiangsu Open Univ, Sch Informat Technol, Nanjing, Jiangsu, Peoples R China
[2] Nanjing Normal Univ, Sch Comp & Elect Informat, Nanjing, Jiangsu, Peoples R China
[3] Nanjing Univ Informat Sci & Technol, Sch Artificial Intelligence, Sch Future Technol, Nanjing, Jiangsu, Peoples R China
[4] Kogakuin Univ, Grad Sch Engn, Dept Elect Engn & Elect, Tokyo, Japan
来源
PLOS ONE | 2024年 / 19卷 / 12期
基金
中国国家自然科学基金;
关键词
D O I
10.1371/journal.pone.0309841
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Deep neural networks have powerful memory capabilities, yet they frequently suffer from overfitting to noisy labels, leading to a decline in classification and generalization performance. To address this issue, sample selection methods that filter out potentially clean labels have been proposed. However, there is a significant gap in size between the filtered, possibly clean subset and the unlabeled subset, which becomes particularly pronounced at high-noise rates. Consequently, this results in underutilizing label-free samples in sample selection methods, leaving room for performance improvement. This study introduces an enhanced sample selection framework with an oversampling strategy (SOS) to overcome this limitation. This framework leverages the valuable information contained in label-free instances to enhance model performance by combining an SOS with state-of-the-art sample selection methods. We validate the effectiveness of SOS through extensive experiments conducted on both synthetic noisy datasets and real-world datasets such as CIFAR, WebVision, and Clothing1M. The source code for SOS will be made available at https://github.com/LanXiaoPang613/SOS.
引用
收藏
页数:37
相关论文
共 50 条
  • [1] Solving Continual Learning with Noisy Labels by Sample Selection and Replay
    Luo, Yiwei
    Jiang, Min
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [2] No regret sample selection with noisy labels
    Song, Heon
    Mitsuo, Nariaki
    Uchida, Seiichi
    Suehiro, Daiki
    MACHINE LEARNING, 2024, 113 (03) : 1163 - 1188
  • [3] No regret sample selection with noisy labels
    Heon Song
    Nariaki Mitsuo
    Seiichi Uchida
    Daiki Suehiro
    Machine Learning, 2024, 113 : 1163 - 1188
  • [4] Sample Selection Approach with Number of False Predictions for Learning with Noisy Labels
    Nomura, Yuichiro
    Kurita, Takio
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (10) : 1759 - 1768
  • [5] Learning with noisy labels using collaborative sample selection and contrastive semi
    Miao, Qing
    Wu, Xiaohe
    Xu, Chao
    Ji, Yanli
    Zuo, Wangmeng
    Guo, Yiwen
    Meng, Zhaopeng
    KNOWLEDGE-BASED SYSTEMS, 2024, 296
  • [6] Knockoffs-SPR: Clean Sample Selection in Learning With Noisy Labels
    Wang, Yikai
    Fu, Yanwei
    Sun, Xinwei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (05) : 3242 - 3256
  • [7] PSSCL: A progressive sample selection framework with contrastive loss designed for noisy labels
    Zhang, Qian
    Zhu, Yi
    Cordeiro, Filipe R.
    Chen, Qiu
    PATTERN RECOGNITION, 2025, 161
  • [8] Joint Optimization Framework for Learning with Noisy Labels
    Tanaka, Daiki
    Ikami, Daiki
    Yamasaki, Toshihiko
    Aizawa, Kiyoharu
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 5552 - 5560
  • [9] A Framework Using Contrastive Learning for Classification with Noisy Labels
    Ciortan, Madalina
    Dupuis, Romain
    Peel, Thomas
    DATA, 2021, 6 (06)
  • [10] A fundus image classification framework for learning with noisy labels
    Hu, Tingxin
    Yang, Bingyu
    Guo, Jia
    Zhang, Weihang
    Liu, Hanruo
    Wang, Ningli
    Li, Huiqi
    COMPUTERIZED MEDICAL IMAGING AND GRAPHICS, 2023, 108