Self-Supervised Modality-Aware Multiple Granularity Pre-Training for RGB-Infrared Person Re-Identification

被引:5
|
作者
Wan, Lin [1 ]
Jing, Qianyan [1 ]
Sun, Zongyuan [1 ]
Zhang, Chuang [2 ]
Li, Zhihang [3 ]
Chen, Yehansen [1 ]
机构
[1] China Univ Geosci, Sch Comp Sci, Wuhan 430078, Peoples R China
[2] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
[3] Chinese Acad Sci, Inst Automat, Beijing 100190, Peoples R China
关键词
Task analysis; Training; Feature extraction; Lighting; Cameras; Visualization; Self-supervised learning; Cross-modality person re-identification; self-supervised learning; multi-modality pre-training;
D O I
10.1109/TIFS.2023.3273911
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
RGB-Infrared person re-identification (RGB-IR ReID) aims to associate people across disjoint RGB and IR camera views. Currently, state-of-the-art performance of RGB-IR ReID is not as impressive as that of conventional ReID. Much of that is due to the notorious modality bias training issue brought by the single-modality ImageNet pre-training, which might yield RGB-biased representations that severely hinder the cross-modality image retrieval. This paper makes first attempt to tackle the task from a pre-training perspective. We propose a self-supervised pre-training solution, named Modality-Aware Multiple Granularity Learning (MMGL), which directly trains models from scratch only on multi-modal ReID datasets, but achieving competitive results against ImageNet pre-training, without using any external data or sophisticated tuning tricks. First, we develop a simple-but-effective 'permutation recovery' pretext task that globally maps shuffled RGB-IR images into a shared latent permutation space, providing modality-invariant global representations for downstream ReID tasks. Second, we present a part-aware cycle-contrastive (PCC) learning strategy that utilizes cross-modality cycle-consistency to maximize agreement between semantically similar RGB-IR image patches. This enables contrastive learning for the unpaired multi-modal scenarios, further improving the discriminability of local features without laborious instance augmentation. Based on these designs, MMGL effectively alleviates the modality bias training problem. Extensive experiments demonstrate that it learns better representations (+8.03% Rank-1 accuracy) with faster training speed (converge only in few hours) and higher data efficiency (< 5% data size) than ImageNet pre-training. The results also suggest it generalizes well to various existing models, losses and has promising transferability across datasets. The code will be released at https://github.com/hansonchen1996/MMGL.
引用
收藏
页码:3044 / 3057
页数:14
相关论文
共 50 条
  • [21] Unsupervised Pre-training for Person Re-identification
    Fu, Dengpan
    Chen, Dongdong
    Bao, Jianmin
    Yang, Hao
    Yuan, Lu
    Zhang, Lei
    Li, Houqiang
    Chen, Dong
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 14745 - 14754
  • [22] Cross-modality paired-images generation and augmentation for RGB-infrared person re-identification
    Wang, Guan'an
    Yang, Yang
    Zhang, Tianzhu
    Cheng, Jian
    Hou, Zengguang
    Tiwari, Prayag
    Pandey, Hari Mohan
    NEURAL NETWORKS, 2020, 128 : 294 - 304
  • [23] RGB-Infrared Cross-Modality Person Re-Identification via Joint Pixel and Feature Alignment
    Wang, Guan'an
    Zhang, Tianzhu
    Cheng, Jian
    Liu, Si
    Yang, Yang
    Hou, Zengguang
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 3622 - 3631
  • [24] RGB-INFRARED PERSON RE-IDENTIFICATION VIA MULTI-MODALITY RELATION AGGREGATION AND GRAPH CONVOLUTION NETWORK
    Sun, Jiangshan
    Zhang, Taiping
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 1174 - 1178
  • [25] Homogeneous-to-Heterogeneous: Unsupervised Learning for RGB-Infrared Person Re-Identification
    Liang, Wenqi
    Wang, Guangcong
    Lai, Jianhuang
    Xie, Xiaohua
    IEEE Transactions on Image Processing, 2021, 30 : 6392 - 6407
  • [26] Proxy-Based Embedding Alignment for RGB-Infrared Person Re-Identification
    Dou, Zhaopeng
    Sun, Yifan
    Li, Yali
    Wang, Shengjin
    TSINGHUA SCIENCE AND TECHNOLOGY, 2025, 30 (03): : 1112 - 1124
  • [27] Two-way constraint network for RGB-Infrared person re-identification
    Zeng, Haitang
    Hu, Weipeng
    Chen, Dihu
    Hu, Haifeng
    ELECTRONICS LETTERS, 2021, 57 (17) : 653 - 655
  • [28] Cross-Modality Person Re-Identification via Modality-Aware Collaborative Ensemble Learning
    Ye, Mang
    Lan, Xiangyuan
    Leng, Qingming
    Shen, Jianbing
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 (29) : 9387 - 9399
  • [29] Homogeneous-to-Heterogeneous: Unsupervised Learning for RGB-Infrared Person Re-Identification
    Liang, Wenqi
    Wang, Guangcong
    Lai, Jianhuang
    Xie, Xiaohua
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 6392 - 6407
  • [30] Self-supervised recalibration network for person re-identification
    Hou, Shaoqi
    Wang, Zhiming
    Dong, Zhihua
    Li, Ye
    Wang, Zhiguo
    Yin, Guangqiang
    Wang, Xinzhong
    DEFENCE TECHNOLOGY, 2024, 31 : 163 - 178