ANNE: Adaptive Nearest Neighbours and Eigenvector-based sample selection for robust learning with noisy labels

被引:1
|
作者
Cordeiro, Filipe R. [1 ]
Carneiro, Gustavo [2 ]
机构
[1] Univ Fed Rural Pernambuco, Dept Comp, Visual Comp Lab, Recife, Brazil
[2] Univ Surrey, Ctr Vis Speech & Signal Proc, Guildford, England
基金
英国工程与自然科学研究理事会;
关键词
Noisy label learning; Deep learning; Sample selection;
D O I
10.1016/j.patcog.2024.111132
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An important stage of most state-of-the-art (SOTA) noisy-label learning methods consists of a sample selection procedure that classifies samples from the noisy-label training set into noisy-label or clean-label subsets. The process of sample selection typically consists of one of the two approaches: loss-based sampling, where high-loss samples are considered to have noisy labels, or feature-based sampling, where samples from the same class tend to cluster together in the feature space and noisy-label samples are identified as anomalies within those clusters. Empirically, loss-based sampling is robust to a wide range of noise rates, while feature-based sampling tends to work effectively in particular scenarios, e.g., the filtering of noisy instances via their eigenvectors (FINE) sampling exhibits greater robustness in scenarios with low noise rates, and the K nearest neighbour (KNN) sampling mitigates better high noise-rate problems. This paper introduces the Adaptive Nearest Neighbours and Eigenvector-based (ANNE) sample selection methodology, a novel approach that integrates loss-based sampling with the feature-based sampling methods FINE and Adaptive KNN to optimize performance across a wide range of noise rate scenarios. ANNE achieves this integration by first partitioning the training set into high-loss and low-loss sub-groups using loss-based sampling. Subsequently, within the low-loss subset, sample selection is performed using FINE, while the high-loss subset employs Adaptive KNN for effective sample selection. We integrate ANNE into the noisy-label learning state of the art (SOTA) method SSR+, and test it on CIFAR-10/-100 (with symmetric, asymmetric and instance-dependent noise), Webvision and ANIMAL-10, where our method shows better accuracy than the SOTA inmost experiments, with a competitive training time. The code is available at https://github.com/filipe-research/anne.
引用
收藏
页数:10
相关论文
共 28 条
  • [1] An improved sample selection framework for learning with noisy labels
    Zhang, Qian
    Zhu, Yi
    Yang, Ming
    Jin, Ge
    Zhu, Yingwen
    Lu, Yanjun
    Zou, Yu
    Chen, Qiu
    PLOS ONE, 2024, 19 (12):
  • [2] Training Robust Deep Neural Networks on Noisy Labels Using Adaptive Sample Selection With Disagreement
    Takeda, Hiroshi
    Yoshida, Soh
    Muneyasu, Mitsuji
    IEEE ACCESS, 2021, 9 : 141131 - 141143
  • [3] Penalty based robust learning with noisy labels
    Kong, Kyeongbo
    Lee, Junggi
    Kwak, Youngchul
    Cho, Young-Rae
    Kim, Seong-Eun
    Song, Woo-Jin
    NEUROCOMPUTING, 2022, 489 : 112 - 127
  • [4] Solving Continual Learning with Noisy Labels by Sample Selection and Replay
    Luo, Yiwei
    Jiang, Min
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [5] Sample Prior Guided Robust Model Learning to Suppress Noisy Labels
    Chen, Wenkai
    Zhu, Chuang
    Li, Mengting
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT II, 2023, 14170 : 3 - 19
  • [6] Sample Selection Approach with Number of False Predictions for Learning with Noisy Labels
    Nomura, Yuichiro
    Kurita, Takio
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (10) : 1759 - 1768
  • [7] Learning with noisy labels using collaborative sample selection and contrastive semi
    Miao, Qing
    Wu, Xiaohe
    Xu, Chao
    Ji, Yanli
    Zuo, Wangmeng
    Guo, Yiwen
    Meng, Zhaopeng
    KNOWLEDGE-BASED SYSTEMS, 2024, 296
  • [8] Knockoffs-SPR: Clean Sample Selection in Learning With Noisy Labels
    Wang, Yikai
    Fu, Yanwei
    Sun, Xinwei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (05) : 3242 - 3256
  • [9] Learning with Noisy Labels by Adaptive Gradient-Based Outlier Removal
    Sedova, Anastasiia
    Zellinger, Lena
    Roth, Benjamin
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT I, 2023, 14169 : 237 - 253
  • [10] Adaptive Sample Selection for Robust Learning under Label Noise
    Patel, Deep
    Sastry, P. S.
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 3921 - 3931