Radar gait recognition using Dual-branch Swin Transformer with Asymmetric Attention Fusion

被引:0
|
作者
He, Wentao [1 ,2 ]
Ren, Jianfeng [2 ,3 ]
Bai, Ruibin [2 ,3 ]
Jiang, Xudong [4 ]
机构
[1] Ningbo Univ, Fac Elect Engn & Comp Sci, Ningbo 315211, Peoples R China
[2] Univ Nottingham Ningbo China, Sch Comp Sci, Digital Port Technol Lab, 199 Taikang East Rd, Ningbo 315100, Peoples R China
[3] Univ Nottingham Ningbo China, Nottingham Ningbo China Beacons Excellence Res & I, 199 Taikang East Rd, Ningbo 315100, Peoples R China
[4] Nanyang Technol Univ, Sch Elect & Elect Engn, 50 Nanyang Ave, Singapore City 639798, Singapore
基金
中国国家自然科学基金;
关键词
Micro-Doppler signature; Radar gait recognition; Spectrogram; Cadence velocity diagram; Asymmetric Attention Fusion; IDENTIFICATION; NETWORKS; IMAGE;
D O I
10.1016/j.patcog.2024.111101
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video-based gait recognition suffers from potential privacy issues and performance degradation due to dim environments, partial occlusions, or camera view changes. Radar has recently become increasingly popular and overcome various challenges presented by vision sensors. To capture tiny differences in radar gait signatures of different people, a dual-branch Swin Transformer is proposed, where one branch captures the time variations of the radar micro-Doppler signature and the other captures the repetitive frequency patterns in the spectrogram. Unlike natural images where objects can be translated, rotated, or scaled, the spatial coordinates of spectrograms and CVDs have unique physical meanings, and there is no affine transformation for radar targets in these synthetic images. The patch splitting mechanism in Vision Transformer makes it ideal to extract discriminant information from patches, and learn the attentive information across patches, as each patch carries some unique physical properties of radar targets. Swin Transformer consists of a set of cascaded Swin blocks to extract semantic features from shallow to deep representations, further improving the classification performance. Lastly, to highlight the branch with larger discriminant power, an Asymmetric Attention Fusion is proposed to optimally fuse the discriminant features from the two branches. To enrich the research on radar gait recognition, a large-scale NTU-RGR dataset is constructed, containing 45,768 radar frames of 98 subjects. The proposed method is evaluated on the NTU-RGR dataset and the MMRGait-1.0 database. It consistently and significantly outperforms all the compared methods on both datasets. The codes are available at: https://github.com/wentaoheunnc/NTU-RGR.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] Swin-Fusion: Swin-Transformer with Feature Fusion for Human Action Recognition
    Chen, Tiansheng
    Mo, Lingfei
    NEURAL PROCESSING LETTERS, 2023, 55 (08) : 11109 - 11130
  • [32] Swin-Fusion: Swin-Transformer with Feature Fusion for Human Action Recognition
    Tiansheng Chen
    Lingfei Mo
    Neural Processing Letters, 2023, 55 : 11109 - 11130
  • [33] RFTNet: Region-Attention Fusion Network Combined with Dual-Branch Vision Transformer for Multimodal Brain Tumor Image Segmentation
    Jiao, Chunxia
    Yang, Tiejun
    Yan, Yanghui
    Yang, Aolin
    ELECTRONICS, 2024, 13 (01)
  • [34] Real-time microexpression recognition in educational scenarios using a dual-branch continuous attention network
    Lv, Yan
    Ning, Meng
    Zhou, Fan
    Lv, Pengfei
    Zhang, Peiying
    Wang, Jian
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (19): : 27123 - 27146
  • [35] A Nonintrusive Load Identification Method Based on Dual-Branch Attention GRU Fusion Network
    Yuan, Jie
    Jin, Ran
    Wang, Lidong
    Wang, Ting
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73
  • [36] DDT: Dual-branch Deformable Transformer for Image Denoising
    Liu, Kangliang
    Du, Xiangcheng
    Liu, Sijie
    Zheng, Yingbin
    Wu, Xingjiao
    Jin, Cheng
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 2765 - 2770
  • [37] GaitFFDA: Feature Fusion and Dual Attention Gait Recognition Model
    Wu, Zhixiong
    Cui, Yong
    TSINGHUA SCIENCE AND TECHNOLOGY, 2025, 30 (01): : 345 - 356
  • [38] Dual-branch channel attention enhancement feature fusion network for diabetic retinopathy segmentation
    Ma, Lei
    Liu, Ziqian
    Xu, Qihang
    Hong, Hanyu
    Wang, Lei
    Zhu, Ying
    Shi, Yu
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 106
  • [39] Dual-Branch Collaborative Transformer for Virtual Try-On
    Fenocchi, Emanuele
    Morelli, Davide
    Cornia, Marcella
    Baraldi, Lorenzo
    Cesari, Fabio
    Cucchiara, Rita
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 2246 - 2250
  • [40] A Novel Dual-Branch Transformer With Gated Cross Attention for Remaining Useful Life Prediction of Bearings
    Cui, Jin
    Ji, J. C.
    Zhang, Tianxiao
    Cao, Licai
    Chen, Zixu
    Ni, Qing
    IEEE SENSORS JOURNAL, 2024, 24 (24) : 41410 - 41423