To Aggregate or Not? Learning with Separate Noisy Labels

被引:10
|
作者
Wei, Jiaheng [1 ]
Zhu, Zhaowei [1 ]
Luo, Tianyi [2 ]
Amid, Ehsan [3 ]
Kumar, Abhishek [3 ]
Liu, Yang [1 ]
机构
[1] Univ Calif Santa Cruz, Santa Cruz, CA 95064 USA
[2] Amazon Search Sci & AI, Palo Alto, CA USA
[3] Google Res, Brain Team, Mountain View, CA USA
基金
美国国家科学基金会;
关键词
Crowdsourcing; Label Aggregation; Label Noise; Human Annotation; LOWER BOUNDS; MODELS;
D O I
10.1145/3580305.3599522
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The rawly collected training data often comes with separate noisy labels collected from multiple imperfect annotators (e.g., via crowdsourcing). A typical way of using these separate labels is to first aggregate them into one and apply standard training methods. The literature has also studied extensively on effective aggregation approaches. This paper revisits this choice and aims to provide an answer to the question of whether one should aggregate separate noisy labels into single ones or use them separately as given. We theoretically analyze the performance of both approaches under the empirical risk minimization framework for a number of popular loss functions, including the ones designed specifically for the problem of learning with noisy labels. Our theorems conclude that label separation is preferred over label aggregation when the noise rates are high, or the number of labelers/annotations is insufficient. Extensive empirical results validate our conclusions.
引用
收藏
页码:2523 / 2535
页数:13
相关论文
共 50 条
  • [41] MILD: Modeling the Instance Learning Dynamics for Learning with Noisy Labels
    Hu, Chuanyang
    Yan, Shipeng
    Gao, Zhitong
    He, Xuming
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 828 - 836
  • [42] Labeling Chaos to Learning Harmony: Federated Learning with Noisy Labels
    Tsouvalas, Vasileios
    Saeed, Aaqib
    Ozcelebi, Tanir
    Meratnia, Nirvana
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2024, 15 (02) : 1 - 26
  • [43] Supervised contrastive learning with corrected labels for noisy label learning
    Jihong Ouyang
    Chenyang Lu
    Bing Wang
    Changchun Li
    Applied Intelligence, 2023, 53 : 29378 - 29392
  • [44] Supervised contrastive learning with corrected labels for noisy label learning
    Ouyang, Jihong
    Lu, Chenyang
    Wang, Bing
    Li, Changchun
    APPLIED INTELLIGENCE, 2023, 53 (23) : 29378 - 29392
  • [45] FedELR: When federated learning meets learning with noisy labels
    Pu, Ruizhi
    Yu, Lixing
    Xu, Gezheng
    Zhou, Fan
    Ling, Charles X.
    Wang, Boyu
    NEURAL NETWORKS, 2025, 187
  • [46] MetaLabelNet: Learning to Generate Soft-Labels From Noisy-Labels
    Algan, Gorkem
    Ulusoy, Ilkay
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 4352 - 4362
  • [47] Learning with Noisy labels via Self-supervised Adversarial Noisy Masking
    Tu, Yuanpeng
    Zhang, Boshen
    Li, Yuxi
    Liu, Liang
    Li, Jian
    Zhang, Jiangning
    Wang, Yabiao
    Wang, Chengjie
    Zhao, Cai Rong
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 16186 - 16195
  • [48] SplitNet: Learnable Clean-Noisy Label Splitting for Learning with Noisy Labels
    Kim, Daehwan
    Ryoo, Kwangrok
    Cho, Hansang
    Kim, Seungryong
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025, 133 (02) : 549 - 566
  • [49] Deep Self-Learning From Noisy Labels
    Han, Jiangfan
    Luo, Ping
    Wang, Xiaogang
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 5137 - 5146
  • [50] Learning Locality Maps from Noisy Geospatial Labels
    Dahiya, Manjeet
    Samatia, Devendra
    Rustogi, Kabir
    PROCEEDINGS OF THE 35TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING (SAC'20), 2020, : 601 - 608