Cross-modal pedestrian re-recognition based on attention mechanism

被引:3
|
作者
Zhao, Yuyao [1 ]
Zhou, Hang [1 ]
Cheng, Hai [1 ]
Huang, Chunguang [1 ]
机构
[1] Heilongjiang Univ, Coll Elect & Engn, Harbin, Heilongjiang, Peoples R China
来源
VISUAL COMPUTER | 2024年 / 40卷 / 04期
基金
中国国家自然科学基金;
关键词
Person re-identification; Cross-modality; Attention mechanisms; Data integration;
D O I
10.1007/s00371-023-02926-7
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Person re-identification, as an essential research direction in intelligent security, has gained the focus of researchers and scholars. In practical scenarios, visible light cameras depend highly on lighting conditions and have limited detection capability in poor light. Therefore, many scholars have gradually shifted their research goals to cross-modality person re-identification. However, there are few relevant studies, and challenges remain in resolving the differences in the images of different modalities. In order to solve these problems, this paper will use the research method based on the attention mechanism to narrow the difference between the two modes and guide the network in a more appropriate direction to improve the recognition performance of the network. Aiming at the problem of using the attention mechanism method can improve training efficiency. However, it is easy to cause the model training instability. This paper proposes a cross-modal pedestrian re-recognition method based on the attention mechanism. A new attention mechanism module is designed to allow the network to use less time to focus on more critical features of a person. In addition, a cross-modality hard center triplet loss is designed to supervise the model training better. The paper has conducted extensive experiments on the above two methods on two publicly available datasets, which obtained better performance than similar current methods and verified the effectiveness and feasibility of the proposed methods in this paper.
引用
收藏
页码:2405 / 2418
页数:14
相关论文
共 50 条
  • [31] Auditory Attention Detection via Cross-Modal Attention
    Cai, Siqi
    Li, Peiwen
    Su, Enze
    Xie, Longhan
    FRONTIERS IN NEUROSCIENCE, 2021, 15
  • [32] Pedestrian Re-Recognition Algorithm Based on Optimization Deep Learning-Sequence Memory Model
    An, Feng-Ping
    COMPLEXITY, 2019, 2019
  • [33] Cross-modal body representation based on visual attention by saliency
    Hikita, Mai
    Fuke, Sawa
    Ogino, Masaki
    Asada, Minoru
    2008 IEEE/RSJ INTERNATIONAL CONFERENCE ON ROBOTS AND INTELLIGENT SYSTEMS, VOLS 1-3, CONFERENCE PROCEEDINGS, 2008, : 2041 - +
  • [34] Emotion recognition using cross-modal attention from EEG and facial expression
    Cui, Rongxuan
    Chen, Wanzhong
    Li, Mingyang
    KNOWLEDGE-BASED SYSTEMS, 2024, 304
  • [35] Cross-Modal Learning with 3D Deformable Attention for Action Recognition
    Kim, Sangwon
    Ahn, Dasom
    Ko, Byoung Chul
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 10231 - 10241
  • [36] Scattering and Optical Cross-Modal Attention Distillation Framework for SAR Target Recognition
    Wang, Longfei
    Liu, Zhunga
    Zhang, Zuowei
    IEEE SENSORS JOURNAL, 2025, 25 (02) : 3126 - 3137
  • [37] PedFormer: Pedestrian Behavior Prediction via Cross-Modal Attention Modulation and Gated Multitask Learning
    Rasouli, Amir
    Kotseruba, Iuliia
    arXiv, 2022,
  • [38] PedFormer: Pedestrian Behavior Prediction via Cross-Modal Attention Modulation and Gated Multitask Learning
    Rasouli, Amir
    Kotseruba, Iuliia
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 9844 - 9851
  • [39] HAPTIC AND CROSS-MODAL RECOGNITION IN CHILDREN
    BUSHNELL, EW
    BULLETIN OF THE PSYCHONOMIC SOCIETY, 1991, 29 (06) : 499 - 499
  • [40] Cross-Modal Distillation for Speaker Recognition
    Jin, Yufeng
    Hu, Guosheng
    Chen, Haonan
    Miao, Duoqian
    Hu, Liang
    Zhao, Cairong
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 12977 - 12985