Information disentanglement based cross-modal representation learning for visible-infrared person re-identification

被引:1
|
作者
Zhu, Xiaoke [1 ]
Zheng, Minghao [1 ]
Chen, Xiaopan [2 ]
Zhang, Xinyu [3 ]
Yuan, Caihong [1 ]
Zhang, Fan [1 ,4 ]
机构
[1] Henan Univ, Sch Comp & Informat Engn, Kaifeng, Peoples R China
[2] Henan Univ, Henan Key Lab Big Data Anal & Proc, Kaifeng, Peoples R China
[3] Wuhan Univ, Sch Comp, Wuhan, Peoples R China
[4] Henan Univ, Henan Engn Res Ctr Intelligent Technol & Applicat, Kaifeng, Peoples R China
基金
中国国家自然科学基金;
关键词
Cross-modal feature learning; Information disentanglement; Shared and specific feature learning; Visible-infrared person re-identification;
D O I
10.1007/s11042-022-13669-3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Visible-infrared person re-identification (VI-ReID) is an important but very challenging task in the automated video surveillance and forensics. Although existing VI-ReID methods have achieved very encouraging results, how to make full use of the useful information contained in cross-modality visible and infrared images has not been well studied. In this paper, we propose an Information Disentanglement based Cross-modal Representation Learning (IDCRL) approach for VI-ReID. Specifically, IDCRL first extracts the shared and specific features from data of each modality by using the shared feature learning module and the specific feature learning module, respectively. To ensure that the shared and specific information can be well disentangled, we impose an orthogonality constraint on the shared and specific features of each modality. To make the shared features extracted from the visible and infrared images of the same person own high similarity, IDCRL designs a shared feature consistency constraint. Furthermore, IDCRL uses a modality-aware loss to ensure that the useful modality-specific features can be extracted from each modality effectively. Then, the obtained shared and specific features are concatenated as the representation of each image. Finally, identity loss function and cross-modal discriminant loss function are employed to enhance the discriminability of the obtained image representation. We conducted comprehensive experiments on the benchmark visible-infrared pedestrian datasets (SYSU-MM01 and RegDB) to evaluate the efficacy of our IDCRL approach. Experimental results demonstrate that IDCRL outperforms the compared state-of-the-art methods. On the SYSU-MM01 dataset, the rank-1 matching rate of our approach reaches 62.35% and 71.64% in the all-search and in-door modes, respectively. On the RegDB dataset, the rank-1 result of our approach reaches 76.32% and 75.49% in the visible to thermal and thermal to visible modes, respectively.
引用
收藏
页码:37983 / 38009
页数:27
相关论文
共 50 条
  • [21] SDL: Spectrum-Disentangled Representation Learning for Visible-Infrared Person Re-Identification
    Kansal, Kajal
    Subramanyam, A. V.
    Wang, Zheng
    Satoh, Shin'ichi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (10) : 3422 - 3432
  • [22] Advancing Visible-Infrared Person Re-Identification: Synergizing Visual-Textual Reasoning and Cross-Modal Feature Alignment
    Qiu, Yuxuan
    Wang, Liyang
    Song, Wei
    Liu, Jiawei
    Shi, Zhiping
    Jiang, Na
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2025, 20 : 2184 - 2196
  • [23] Auxiliary Representation Guided Network for Visible-Infrared Person Re-Identification
    Qi, Mengzan
    Chan, Sixian
    Hang, Chen
    Zhang, Guixu
    Zeng, Tieyong
    Li, Zhi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 340 - 355
  • [24] Visible-infrared person re-identification model based on feature consistency and modal indistinguishability
    Sun, Jia
    Li, Yanfeng
    Chen, Houjin
    Peng, Yahui
    Zhu, Jinlei
    MACHINE VISION AND APPLICATIONS, 2023, 34 (01)
  • [25] Visible-infrared person re-identification model based on feature consistency and modal indistinguishability
    Jia Sun
    Yanfeng Li
    Houjin Chen
    Yahui Peng
    Jinlei Zhu
    Machine Vision and Applications, 2023, 34
  • [26] Visible-Infrared Person Re-Identification via Homogeneous Augmented Tri-Modal Learning
    Ye, Mang
    Shen, Jianbing
    Shao, Ling
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 : 728 - 739
  • [27] Cross-Modality Semantic Consistency Learning for Visible-Infrared Person Re-Identification
    Liu, Min
    Zhang, Zhu
    Bian, Yuan
    Wang, Xueping
    Sun, Yeqing
    Zhang, Baida
    Wang, Yaonan
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 568 - 580
  • [28] Dual-Adversarial Representation Disentanglement for Visible Infrared Person Re-Identification
    Wei, Ziyu
    Yang, Xi
    Wang, Nannan
    Gao, Xinbo
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 2186 - 2200
  • [29] An efficient framework for visible-infrared cross modality person re-identification
    Basaran, Emrah
    Gokmen, Muhittin
    Kamasak, Mustafa E.
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2020, 87
  • [30] Cross-Modality Transformer for Visible-Infrared Person Re-Identification
    Jiang, Kongzhu
    Zhang, Tianzhu
    Liu, Xiang
    Qian, Bingqiao
    Zhang, Yongdong
    Wu, Feng
    COMPUTER VISION - ECCV 2022, PT XIV, 2022, 13674 : 480 - 496