Self-Supervised Modality-Aware Multiple Granularity Pre-Training for RGB-Infrared Person Re-Identification

被引:5
|
作者
Wan, Lin [1 ]
Jing, Qianyan [1 ]
Sun, Zongyuan [1 ]
Zhang, Chuang [2 ]
Li, Zhihang [3 ]
Chen, Yehansen [1 ]
机构
[1] China Univ Geosci, Sch Comp Sci, Wuhan 430078, Peoples R China
[2] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
[3] Chinese Acad Sci, Inst Automat, Beijing 100190, Peoples R China
关键词
Task analysis; Training; Feature extraction; Lighting; Cameras; Visualization; Self-supervised learning; Cross-modality person re-identification; self-supervised learning; multi-modality pre-training;
D O I
10.1109/TIFS.2023.3273911
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
RGB-Infrared person re-identification (RGB-IR ReID) aims to associate people across disjoint RGB and IR camera views. Currently, state-of-the-art performance of RGB-IR ReID is not as impressive as that of conventional ReID. Much of that is due to the notorious modality bias training issue brought by the single-modality ImageNet pre-training, which might yield RGB-biased representations that severely hinder the cross-modality image retrieval. This paper makes first attempt to tackle the task from a pre-training perspective. We propose a self-supervised pre-training solution, named Modality-Aware Multiple Granularity Learning (MMGL), which directly trains models from scratch only on multi-modal ReID datasets, but achieving competitive results against ImageNet pre-training, without using any external data or sophisticated tuning tricks. First, we develop a simple-but-effective 'permutation recovery' pretext task that globally maps shuffled RGB-IR images into a shared latent permutation space, providing modality-invariant global representations for downstream ReID tasks. Second, we present a part-aware cycle-contrastive (PCC) learning strategy that utilizes cross-modality cycle-consistency to maximize agreement between semantically similar RGB-IR image patches. This enables contrastive learning for the unpaired multi-modal scenarios, further improving the discriminability of local features without laborious instance augmentation. Based on these designs, MMGL effectively alleviates the modality bias training problem. Extensive experiments demonstrate that it learns better representations (+8.03% Rank-1 accuracy) with faster training speed (converge only in few hours) and higher data efficiency (< 5% data size) than ImageNet pre-training. The results also suggest it generalizes well to various existing models, losses and has promising transferability across datasets. The code will be released at https://github.com/hansonchen1996/MMGL.
引用
收藏
页码:3044 / 3057
页数:14
相关论文
共 50 条
  • [1] Modality-aware Style Adaptation for RGB-Infrared Person Re-Identification
    Miao, Ziling
    Liu, Hong
    Shi, Wei
    Xu, Wanlu
    Ye, Hanrong
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 916 - 922
  • [2] PASS: Part-Aware Self-Supervised Pre-Training for Person Re-Identification
    Zhu, Kuan
    Guo, Haiyun
    Yan, Tianyi
    Zhu, Yousong
    Wang, Jinqiao
    Tang, Ming
    COMPUTER VISION - ECCV 2022, PT XIV, 2022, 13674 : 198 - 214
  • [3] RGB-Infrared Cross-Modality Person Re-Identification
    Wu, Ancong
    Zheng, Wei-Shi
    Yu, Hong-Xing
    Gong, Shaogang
    Lai, Jianhuang
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 5390 - 5399
  • [4] RGB-Infrared Person Re-identification via Image Modality Conversion
    Dai, Huangpeng
    Xie, Qing
    Ma, Yanchun
    Liu, Yongjian
    Xiong, Shengwu
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 592 - 598
  • [5] Self-supervised Pre-training with Learnable Tokenizers for Person Re-Identification in Railway Stations
    Yang, Enze
    Li, Chao
    Liu, Shuoyan
    Liu, Yuxin
    Zhao, Shitao
    Huang, Nan
    2022 16TH IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP2022), VOL 1, 2022, : 325 - 330
  • [6] Deep Constraints Space of Medium Modality for RGB-Infrared Person Re-identification
    Huang, Baojin
    Chen, Hao
    Qin, Wencheng
    NEURAL PROCESSING LETTERS, 2023, 55 (03) : 3007 - 3024
  • [7] Deep Constraints Space of Medium Modality for RGB-Infrared Person Re-identification
    Baojin Huang
    Hao Chen
    Wencheng Qin
    Neural Processing Letters, 2023, 55 : 3007 - 3024
  • [8] Self-Supervised Pre-training on the Target Domain for Cross-Domain Person Re-identification
    Zhang, Junyin
    Ge, Yongxin
    Gu, Xinqian
    Hua, Boyu
    Xiang, Tao
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 4268 - 4276
  • [9] Cross-Modality Channel Mixup and Modality Decorrelation for RGB-Infrared Person Re-Identification
    Hua, Boyu
    Zhang, Junyin
    Li, Ziqiang
    Ge, Yongxin
    IEEE TRANSACTIONS ON BIOMETRICS, BEHAVIOR, AND IDENTITY SCIENCE, 2023, 5 (04): : 512 - 523
  • [10] Deep Attention Network for RGB-Infrared Cross-Modality Person Re-Identification
    Li, Yang
    Xu, Huahu
    4TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE APPLICATIONS AND TECHNOLOGIES (AIAAT 2020), 2020, 1642