To Aggregate or Not? Learning with Separate Noisy Labels

被引:10
|
作者
Wei, Jiaheng [1 ]
Zhu, Zhaowei [1 ]
Luo, Tianyi [2 ]
Amid, Ehsan [3 ]
Kumar, Abhishek [3 ]
Liu, Yang [1 ]
机构
[1] Univ Calif Santa Cruz, Santa Cruz, CA 95064 USA
[2] Amazon Search Sci & AI, Palo Alto, CA USA
[3] Google Res, Brain Team, Mountain View, CA USA
基金
美国国家科学基金会;
关键词
Crowdsourcing; Label Aggregation; Label Noise; Human Annotation; LOWER BOUNDS; MODELS;
D O I
10.1145/3580305.3599522
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The rawly collected training data often comes with separate noisy labels collected from multiple imperfect annotators (e.g., via crowdsourcing). A typical way of using these separate labels is to first aggregate them into one and apply standard training methods. The literature has also studied extensively on effective aggregation approaches. This paper revisits this choice and aims to provide an answer to the question of whether one should aggregate separate noisy labels into single ones or use them separately as given. We theoretically analyze the performance of both approaches under the empirical risk minimization framework for a number of popular loss functions, including the ones designed specifically for the problem of learning with noisy labels. Our theorems conclude that label separation is preferred over label aggregation when the noise rates are high, or the number of labelers/annotations is insufficient. Extensive empirical results validate our conclusions.
引用
收藏
页码:2523 / 2535
页数:13
相关论文
共 50 条
  • [1] Learning to Aggregate and Refine Noisy Labels for Visual Sentiment Analysis
    Zhu, Wei
    Zheng, Zihe
    Zheng, Haitian
    Lyu, Hanjia
    Luo, Jiebo
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 571 - 577
  • [2] Learning and Mining with Noisy Labels
    Sugiyama, Masashi
    Liu, Tongliang
    Han, Bo
    Liu, Yang
    Niu, Gang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 5152 - 5155
  • [3] On Learning Contrastive Representations for Learning with Noisy Labels
    Yi, Li
    Liu, Sheng
    She, Qi
    McLeod, A. Ian
    Wang, Boyu
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 16661 - 16670
  • [4] Learning to rectify for robust learning with noisy labels
    Sun, Haoliang
    Guo, Chenhui
    Wei, Qi
    Han, Zhongyi
    Yin, Yilong
    PATTERN RECOGNITION, 2022, 124
  • [5] Partial Label Learning with Noisy Labels
    Zhao, Pan
    Tang, Long
    Pan, Zhigeng
    Annals of Data Science, 2025, 12 (01) : 199 - 212
  • [6] Augmentation Strategies for Learning with Noisy Labels
    Nishi, Kento
    Ding, Yi
    Rich, Alex
    Hollerer, Tobias
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 8018 - 8027
  • [7] Learning with Neighbor Consistency for Noisy Labels
    Iscen, Ahmet
    Valmadre, Jack
    Arnab, Anurag
    Schmid, Cordelia
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 4662 - 4671
  • [8] DEEP LEARNING CLASSIFICATION WITH NOISY LABELS
    Sanchez, Guillaume
    Guis, Vincente
    Marxer, Ricard
    Bouchara, Frederic
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (ICMEW), 2020,
  • [9] Twin Contrastive Learning with Noisy Labels
    Huang, Zhizhong
    Zhang, Junping
    Shan, Hongming
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11661 - 11670
  • [10] Iterative Cross Learning on Noisy Labels
    Yuan, Bodi
    Chen, Jianyu
    Zhang, Weidong
    Tai, Hung-Shuo
    McMains, Sara
    2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018), 2018, : 757 - 765