Information disentanglement based cross-modal representation learning for visible-infrared person re-identification

被引:1
|
作者
Zhu, Xiaoke [1 ]
Zheng, Minghao [1 ]
Chen, Xiaopan [2 ]
Zhang, Xinyu [3 ]
Yuan, Caihong [1 ]
Zhang, Fan [1 ,4 ]
机构
[1] Henan Univ, Sch Comp & Informat Engn, Kaifeng, Peoples R China
[2] Henan Univ, Henan Key Lab Big Data Anal & Proc, Kaifeng, Peoples R China
[3] Wuhan Univ, Sch Comp, Wuhan, Peoples R China
[4] Henan Univ, Henan Engn Res Ctr Intelligent Technol & Applicat, Kaifeng, Peoples R China
基金
中国国家自然科学基金;
关键词
Cross-modal feature learning; Information disentanglement; Shared and specific feature learning; Visible-infrared person re-identification;
D O I
10.1007/s11042-022-13669-3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Visible-infrared person re-identification (VI-ReID) is an important but very challenging task in the automated video surveillance and forensics. Although existing VI-ReID methods have achieved very encouraging results, how to make full use of the useful information contained in cross-modality visible and infrared images has not been well studied. In this paper, we propose an Information Disentanglement based Cross-modal Representation Learning (IDCRL) approach for VI-ReID. Specifically, IDCRL first extracts the shared and specific features from data of each modality by using the shared feature learning module and the specific feature learning module, respectively. To ensure that the shared and specific information can be well disentangled, we impose an orthogonality constraint on the shared and specific features of each modality. To make the shared features extracted from the visible and infrared images of the same person own high similarity, IDCRL designs a shared feature consistency constraint. Furthermore, IDCRL uses a modality-aware loss to ensure that the useful modality-specific features can be extracted from each modality effectively. Then, the obtained shared and specific features are concatenated as the representation of each image. Finally, identity loss function and cross-modal discriminant loss function are employed to enhance the discriminability of the obtained image representation. We conducted comprehensive experiments on the benchmark visible-infrared pedestrian datasets (SYSU-MM01 and RegDB) to evaluate the efficacy of our IDCRL approach. Experimental results demonstrate that IDCRL outperforms the compared state-of-the-art methods. On the SYSU-MM01 dataset, the rank-1 matching rate of our approach reaches 62.35% and 71.64% in the all-search and in-door modes, respectively. On the RegDB dataset, the rank-1 result of our approach reaches 76.32% and 75.49% in the visible to thermal and thermal to visible modes, respectively.
引用
收藏
页码:37983 / 38009
页数:27
相关论文
共 50 条
  • [1] Information disentanglement based cross-modal representation learning for visible-infrared person re-identification
    Xiaoke Zhu
    Minghao Zheng
    Xiaopan Chen
    Xinyu Zhang
    Caihong Yuan
    Fan Zhang
    Multimedia Tools and Applications, 2023, 82 : 37983 - 38009
  • [2] Cascaded Cross-modal Alignment for Visible-Infrared Person Re-Identification
    Li, Zhaohui
    Wang, Qiangchang
    Chen, Lu
    Zhang, Xinxin
    Yin, Yilong
    KNOWLEDGE-BASED SYSTEMS, 2024, 305
  • [3] Learning by Aligning: Visible-Infrared Person Re-identification using Cross-Modal Correspondences
    Park, Hyunjong
    Lee, Sanghoon
    Lee, Junghyup
    Ham, Bumsub
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 12026 - 12035
  • [4] Progressive Cross-Modal Association Learning for Unsupervised Visible-Infrared Person Re-Identification
    Yang, Yiming
    Hu, Weipeng
    Hu, Haifeng
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2025, 20 : 1290 - 1304
  • [5] Cross-modal group-relation optimization for visible-infrared person re-identification
    Zhu, Jianqing
    Wu, Hanxiao
    Chen, Yutao
    Xu, Heng
    Fu, Yuqing
    Zeng, Huanqiang
    Liu, Liu
    Lei, Zhen
    NEURAL NETWORKS, 2024, 179
  • [6] Identity Feature Disentanglement for Visible-Infrared Person Re-Identification
    Chen, Xiumei
    Zheng, Xiangtao
    Lu, Xiaoqiang
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2023, 19 (06)
  • [7] HCFN: Hierarchical cross-modal shared feature network for visible-infrared person re-identification?
    Li, Yueying
    Zhang, Huaxiang
    Liu, Li
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2022, 89
  • [8] Margin-Based Modal Adaptive Learning for Visible-Infrared Person Re-Identification
    Zhao, Qianqian
    Wu, Hanxiao
    Zhu, Jianqing
    SENSORS, 2023, 23 (03)
  • [9] Multi-knowledge-driven enhanced module for visible-infrared cross-modal person Re-identification
    Shan, Shihao
    Sun, Peixin
    Xiao, Guoqiang
    Wu, Song
    INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2024, 13 (02)
  • [10] Contrastive Learning with Information Compensation for Visible-Infrared Person Re-Identification
    Zhang, La
    Guo, Haiyun
    Zhao, Xu
    Sun, Jian
    Wang, Jinqiao
    2024 14TH ASIAN CONTROL CONFERENCE, ASCC 2024, 2024, : 1266 - 1271