Radar gait recognition using Dual-branch Swin Transformer with Asymmetric Attention Fusion

被引:0
|
作者
He, Wentao [1 ,2 ]
Ren, Jianfeng [2 ,3 ]
Bai, Ruibin [2 ,3 ]
Jiang, Xudong [4 ]
机构
[1] Ningbo Univ, Fac Elect Engn & Comp Sci, Ningbo 315211, Peoples R China
[2] Univ Nottingham Ningbo China, Sch Comp Sci, Digital Port Technol Lab, 199 Taikang East Rd, Ningbo 315100, Peoples R China
[3] Univ Nottingham Ningbo China, Nottingham Ningbo China Beacons Excellence Res & I, 199 Taikang East Rd, Ningbo 315100, Peoples R China
[4] Nanyang Technol Univ, Sch Elect & Elect Engn, 50 Nanyang Ave, Singapore City 639798, Singapore
基金
中国国家自然科学基金;
关键词
Micro-Doppler signature; Radar gait recognition; Spectrogram; Cadence velocity diagram; Asymmetric Attention Fusion; IDENTIFICATION; NETWORKS; IMAGE;
D O I
10.1016/j.patcog.2024.111101
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video-based gait recognition suffers from potential privacy issues and performance degradation due to dim environments, partial occlusions, or camera view changes. Radar has recently become increasingly popular and overcome various challenges presented by vision sensors. To capture tiny differences in radar gait signatures of different people, a dual-branch Swin Transformer is proposed, where one branch captures the time variations of the radar micro-Doppler signature and the other captures the repetitive frequency patterns in the spectrogram. Unlike natural images where objects can be translated, rotated, or scaled, the spatial coordinates of spectrograms and CVDs have unique physical meanings, and there is no affine transformation for radar targets in these synthetic images. The patch splitting mechanism in Vision Transformer makes it ideal to extract discriminant information from patches, and learn the attentive information across patches, as each patch carries some unique physical properties of radar targets. Swin Transformer consists of a set of cascaded Swin blocks to extract semantic features from shallow to deep representations, further improving the classification performance. Lastly, to highlight the branch with larger discriminant power, an Asymmetric Attention Fusion is proposed to optimally fuse the discriminant features from the two branches. To enrich the research on radar gait recognition, a large-scale NTU-RGR dataset is constructed, containing 45,768 radar frames of 98 subjects. The proposed method is evaluated on the NTU-RGR dataset and the MMRGait-1.0 database. It consistently and significantly outperforms all the compared methods on both datasets. The codes are available at: https://github.com/wentaoheunnc/NTU-RGR.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Dual-branch fusion model for lensless imaging
    Zhang, Yinger
    Wu, Zhouyi
    Xu, Yunhui
    Huangfu, Jiangtao
    OPTICS EXPRESS, 2023, 31 (12) : 19463 - 19477
  • [42] Multi-Modal Object Detection Method Based on Dual-Branch Asymmetric Attention Backbone and Feature Fusion Pyramid Network
    Wang, Jinpeng
    Su, Nan
    Zhao, Chunhui
    Yan, Yiming
    Feng, Shou
    REMOTE SENSING, 2024, 16 (20)
  • [43] Masked face recognition via dual-branch convolutional self-attention network
    Wan, Weiguo
    Wen, Runlin
    Deng, Linghan
    Yang, Yong
    APPLIED SOFT COMPUTING, 2025, 169
  • [44] HDCTfusion: Hybrid Dual-Branch Network Based on CNN and Transformer for Infrared and Visible Image Fusion
    Wang, Wenqing
    Li, Lingzhou
    Yang, Yifei
    Liu, Han
    Guo, Runyuan
    SENSORS, 2024, 24 (23)
  • [45] A Dual-Branch Fusion Network Based on Reconstructed Transformer for Building Extraction in Remote Sensing Imagery
    Wang, Yitong
    Wang, Shumin
    Dou, Aixia
    SENSORS, 2024, 24 (02)
  • [46] Gait recognition with global-local feature fusion based on swin transformer-3DCNN
    Wang, Ting
    Zhou, Guanghang
    Pu, Yanfeng
    Moreno, Ramon
    Yang, Guoping
    SIGNAL IMAGE AND VIDEO PROCESSING, 2025, 19 (01)
  • [47] A dual-branch siamese spatial-spectral transformer attention network for Hyperspectral Image Change Detection
    Zhang, Yiyan
    Wang, Tingting
    Zhang, Chenkai
    Xu, Shufang
    Gao, Hongmin
    Li, Chenming
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [48] Group-level Emotion Recognition using Hierarchical Dual-branch Cross Transformer with Semi-supervised Learning
    Xu, Jinke
    Huang, Xiaohua
    2024 IEEE 4TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND ARTIFICIAL INTELLIGENCE, SEAI 2024, 2024, : 252 - 256
  • [49] A Dual-Branch Multiscale Transformer Network for Hyperspectral Image Classification
    Shi, Cuiping
    Yue, Shuheng
    Wang, Liguo
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 20
  • [50] DCFNet: An Effective Dual-Branch Cross-Attention Fusion Network for Medical Image Segmentation
    Zhu, Chengzhang
    Zhang, Renmao
    Xiao, Yalong
    Zou, Beiji
    Chai, Xian
    Yang, Zhangzheng
    Hu, Rong
    Duan, Xuanchu
    CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2024, 140 (01): : 1103 - 1128