Robust Duality Learning for Unsupervised Visible-Infrared Person Re-Identification

被引:0
|
作者
Li, Yongxiang [1 ]
Sun, Yuan [1 ]
Qin, Yang [1 ]
Peng, Dezhong [1 ,2 ]
Peng, Xi [1 ]
Hu, Peng [1 ]
机构
[1] Sichuan Univ, Coll Comp Sci, Chengdu 610065, Peoples R China
[2] Sichuan Natl Innovat New Vis UHD Video Technol Co, Chengdu 610095, Peoples R China
基金
中国国家自然科学基金;
关键词
Noise measurement; Noise; Overfitting; Adaptation models; Training; Predictive models; Semantics; Interference; Robustness; Optimization; Unsupervised VI-ReID; pseudo-label noise; noise correspondence; cluster consistency;
D O I
10.1109/TIFS.2025.3536613
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Unsupervised visible-infrared person re-identification (UVI-ReID) aims at retrieving pedestrian images of the same individual across distinct modalities, presenting challenges due to the inherent heterogeneity gap and the absence of cost-prohibitive annotations. Although existing methods employ self-training with clustering-generated pseudo-labels to bridge this gap, they always implicitly assume that these pseudo-labels are predicted correctly. In practice, however, this presumption is impossible to satisfy due to the difficulty of training a perfect model let alone without any ground truths, resulting in pseudo-labeling errors. Based on the observation, this study introduces a new learning paradigm for UVI-ReID considering Pseudo-Label Noise (PLN), which encompasses three challenges: noise overfitting, error accumulation, and noisy cluster correspondence. To conquer these challenges, we propose a novel robust duality learning framework (RoDE) for UVI-ReID to mitigate the adverse impact of noisy pseudo-labels. Specifically, for noise overfitting, we propose a novel Robust Adaptive Learning mechanism (RAL) to dynamically prioritize clean samples while deprioritizing noisy ones, thus avoiding overemphasizing noise. To circumvent error accumulation of self-training, where the model tends to confirm its mistakes, RoDE alternately trains dual distinct models using pseudo-labels predicted by their counterparts, thereby maintaining diversity and avoiding collapse into noise. However, this will lead to cross-cluster misalignment between the two distinct models, not to mention the misalignment between different modalities, resulting in dual noisy cluster correspondence and thus difficult to optimize. To address this issue, a Cluster Consistency Matching mechanism (CCM) is presented to ensure reliable alignment across distinct modalities as well as across different models by leveraging cross-cluster similarities. Extensive experiments on three benchmark datasets demonstrate the effectiveness of the proposed RoDE.
引用
收藏
页码:1937 / 1948
页数:12
相关论文
共 50 条
  • [31] Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification
    Feng, Zhanxiang
    Lai, Jianhuang
    Xie, Xiaohua
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 579 - 590
  • [32] Learning a Robust Synthetic Modality with Dual-Level Alignment for Visible-Infrared Person Re-identification
    Wang, Zichun
    Cheng, Xu
    PATTERN RECOGNITION AND COMPUTER VISION, PT V, PRCV 2024, 2025, 15035 : 289 - 303
  • [33] Shape-Erased Feature Learning for Visible-Infrared Person Re-Identification
    Feng, Jiawei
    Wu, Ancong
    Zhen, Wei-Shi
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 22752 - 22761
  • [34] Unbiased Feature Learning with Causal Intervention for Visible-Infrared Person Re-Identification
    Yuan, Bo wen
    Lu, Jiahao
    You, Sisi
    Bao, Bing-kun
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 20 (10)
  • [35] Cross-modality consistency learning for visible-infrared person re-identification
    Shao, Jie
    Tang, Lei
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [36] Learning dual attention enhancement feature for visible-infrared person re-identification
    Zhang, Guoqing
    Zhang, Yinyin
    Zhang, Hongwei
    Chen, Yuhao
    Zheng, Yuhui
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 99
  • [37] Multi-Stage Auxiliary Learning for Visible-Infrared Person Re-Identification
    Zhang, Huadong
    Cheng, Shuli
    Du, Anyu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (11) : 12032 - 12047
  • [38] Text-augmented Multi-Modality contrastive learning for unsupervised visible-infrared person re-identification
    Sun, Rui
    Huang, Guoxi
    Wang, Xuebin
    Du, Yun
    Zhang, Xudong
    IMAGE AND VISION COMPUTING, 2024, 152
  • [39] Modality Unifying Network for Visible-Infrared Person Re-Identification
    Yu, Hao
    Cheng, Xu
    Peng, Wei
    Liu, Weihao
    Zhao, Guoying
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11151 - 11161
  • [40] Progressive discrepancy elimination for visible-infrared person re-identification
    Zhang, Guoqing
    Wang, Zhun Zhun
    Wang, Hairui
    Zhou, Jieqiong
    Zheng, Yuhui
    NEUROCOMPUTING, 2024, 607