Transferring Annotator- and Instance-Dependent Transition Matrix for Learning From Crowds

被引:0
|
作者
Li, Shikun [1 ,2 ]
Xia, Xiaobo [3 ]
Deng, Jiankang [4 ]
Ge, Shiming [1 ,2 ]
Liu, Tongliang [3 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing 100095, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing 100049, Peoples R China
[3] Univ Sydney, Fac Engn, Sydney AI Ctr, Sch Comp Sci, Darlington, NSW 2008, Australia
[4] Imperial Coll London, Dept Comp, London SW7 2BX, England
基金
澳大利亚研究理事会;
关键词
Noise; Annotations; Noise measurement; Knowledge transfer; Data models; Sparse matrices; Estimation; Learning from crowds; label-noise learning; noise transition matrix; knowledge transfer;
D O I
10.1109/TPAMI.2024.3388209
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning from crowds describes that the annotations of training data are obtained with crowd-sourcing services. Multiple annotators each complete their own small part of the annotations, where labeling mistakes that depend on annotators occur frequently. Modeling the label-noise generation process by the noise transition matrix is a powerful tool to tackle the label noise. In real-world crowd-sourcing scenarios, noise transition matrices are both annotator- and instance-dependent. However, due to the high complexity of annotator- and instance-dependent transition matrices (AIDTM), annotation sparsity, which means each annotator only labels a tiny part of instances, makes modeling AIDTM very challenging. Without prior knowledge, existing works simplify the problem by assuming the transition matrix is instance-independent or using simple parametric ways, which lose modeling generality. Motivated by this, we target a more realistic problem, estimating general AIDTM in practice. Without losing modeling generality, we parameterize AIDTM with deep neural networks. To alleviate the modeling challenge, we suppose every annotator shares its noise pattern with similar annotators, and estimate AIDTM via knowledge transfer. We hence first model the mixture of noise patterns by all annotators, and then transfer this modeling to individual annotators. Furthermore, considering that the transfer from the mixture of noise patterns to individuals may cause two annotators with highly different noise generations to perturb each other, we employ the knowledge transfer between identified neighboring annotators to calibrate the modeling. Theoretical analyses are derived to demonstrate that both the knowledge transfer from global to individuals and the knowledge transfer between neighboring individuals can effectively help mitigate the challenge of modeling general AIDTM. Experiments confirm the superiority of the proposed approach on synthetic and real-world crowd-sourcing data.
引用
收藏
页码:7377 / 7391
页数:15
相关论文
共 50 条
  • [31] Horizon-Free and Instance-Dependent Regret Bounds for Reinforcement Learning with General Function Approximation
    Huang, Jiayi
    Zhong, Han
    Wang, Liwei
    Yang, Lin F.
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [32] The Hidden Cost of Fraud: An Instance-Dependent Cost-Sensitive Approach for Positive and Unlabeled Learning
    Vazquez, Carlos Ortega
    De Weerdt, Jochen
    vanden Broucke, Seppe
    FOURTH INTERNATIONAL WORKSHOP ON LEARNING WITH IMBALANCED DOMAINS: THEORY AND APPLICATIONS, VOL 183, 2022, 183 : 53 - 67
  • [33] Centrality and Consistency: Two-Stage Clean Samples Identification for Learning with Instance-Dependent Noisy Labels
    Zhao, Ganlong
    Li, Guanbin
    Qin, Yipeng
    Liu, Feng
    Yu, Yizhou
    COMPUTER VISION, ECCV 2022, PT XXV, 2022, 13685 : 21 - 37
  • [34] Instance-Dependent Noisy-Label Learning with Graphical Model Based Noise-Rate Estimation
    Garg, Arpit
    Cuong Nguyen
    Felix, Rafael
    Thanh-Toan Do
    Carneiro, Gustavo
    COMPUTER VISION-ECCV 2024, PT IV, 2025, 15062 : 372 - 389
  • [35] Beyond Value-Function Gaps: Improved Instance-Dependent Regret Bounds for Episodic Reinforcement Learning
    Dann, Chris
    Marinov, Teodor V.
    Mohri, Mehryar
    Zimmert, Julian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [36] Candidate-aware Selective Disambiguation Based On Normalized Entropy for Instance-dependent Partial-label Learning
    He, Shuo
    Yang, Guowu
    Feng, Lei
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 1792 - 1801
  • [37] Tackling Heavy-Tailed Rewards in Reinforcement Learning with Function Approximation: Minimax Optimal and Instance-Dependent Regret Bounds
    Huang, Jiayi
    Zhong, Han
    Wang, Liwei
    Yang, Lin F.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [38] Scaling up the learning-from-crowds GLAD algorithm using instance-difficulty clustering
    Gonzalez Rodrigo, Enrique
    Aledo, Juan A.
    Gamez, Jose A.
    PROGRESS IN ARTIFICIAL INTELLIGENCE, 2019, 8 (03) : 389 - 399
  • [39] Scaling up the learning-from-crowds GLAD algorithm using instance-difficulty clustering
    Enrique González Rodrigo
    Juan A. Aledo
    Jose A. Gamez
    Progress in Artificial Intelligence, 2019, 8 : 389 - 399
  • [40] Learning from Crowds via Joint Probabilistic Matrix Factorization and Clustering in Latent Space
    Yao, Wuguannan
    Lee, Wonjung
    Wang, Junhui
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: APPLIED DATA SCIENCE TRACK, ECML PKDD 2020, PT IV, 2021, 12460 : 546 - 561