Scalable Penalized Regression for Noise Detection in Learning with Noisy Labels

被引:25
|
作者
Wang, Yikai [1 ]
Sun, Xinwei [1 ]
Fu, Yanwei [1 ]
机构
[1] Fudan Univ, Sch Data Sci, Shanghai, Peoples R China
关键词
CONSISTENCY;
D O I
10.1109/CVPR52688.2022.00044
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Noisy training set usually leads to the degradation of generalization and robustness of neural networks. In this paper, we propose using a theoretically guaranteed noisy label detection framework to detect and remove noisy data for Learning with Noisy Labels (LNL). Specifically, we design a penalized regression to model the linear relation between network features and one-hot labels, where the noisy data are identified by the non-zero mean shift parameters solved in the regression model. To make the framework scalable to datasets that contain a large number of categories and training data, we propose a split algorithm to divide the whole training set into small pieces that can be solved by the penalized regression in parallel, leading to the Scalable Penalized Regression (SPR) framework. We provide the non-asymptotic probabilistic condition for SPR to correctly identify the noisy data. While SPR can be regarded as a sample selection module for standard supervised training pipeline, we further combine it with semi-supervised algorithm to further exploit the support of noisy data as unlabeled data. Experimental results on several benchmark datasets and real-world noisy datasets show the effectiveness of our framework.
引用
收藏
页码:346 / 355
页数:10
相关论文
共 50 条
  • [31] FINE Samples for Learning with Noisy Labels
    Kim, Taehyeon
    Ko, Jongwoo
    Cho, Sangwook
    Choi, Jinhwan
    Yun, Se-Young
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [32] Label Distribution for Learning with Noisy Labels
    Liu, Yun-Peng
    Xu, Ning
    Zhang, Yu
    Geng, Xin
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 2568 - 2574
  • [33] Progressive Stochastic Learning for Noisy Labels
    Han, Bo
    Tsang, Ivor W.
    Chen, Ling
    Yu, Celina P.
    Fung, Sai-Fu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (10) : 5136 - 5148
  • [34] Noise-Resilient Federated Learning: Suppressing Noisy Labels in the Local Datasets of Participants
    Mishra, Rahul
    Gupta, Hari Prabhat
    Dutta, Tanima
    IEEE INFOCOM 2022 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (INFOCOM WKSHPS), 2022,
  • [35] Learning From Noisy Labels for MIMO Detection With One-Bit ADCs
    Park, Jinsung
    Lee, Namyoon
    Hong, Song-Nam
    Jeon, Yo-Seb
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2023, 12 (03) : 456 - 460
  • [36] MIMO Detection under Hardware Impairments via Learning from Noisy Labels
    Kwon, Jinman
    Jeon, Yo-Seb
    Poor, H. Vincent
    2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 3192 - 3197
  • [37] LNL plus K: Enhancing Learning with Noisy Labels Through Noise Source Knowledge Integration
    Wang, Siqi
    Plummer, Bryan A.
    COMPUTER VISION - ECCV 2024, PT LXI, 2025, 15119 : 374 - 392
  • [38] A survey on learning with noisy labels in Natural Language Processing: How to train models with label noise
    Zhang, Han
    Zhang, Yazhou
    Li, Jiajun
    Liu, Junxiu
    Ji, Lixia
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 146
  • [39] Learning Noise Transition Matrix from Only Noisy Labels via Total Variation Regularization
    Zhang, Yivan
    Niu, Gang
    Sugiyama, Masashi
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [40] Limited Gradient Descent: Learning With Noisy Labels
    Sun, Yi
    Tian, Yan
    Xu, Yiping
    Li, Jianxiang
    IEEE ACCESS, 2019, 7 : 168296 - 168306