Radar gait recognition using Dual-branch Swin Transformer with Asymmetric Attention Fusion

被引:0
|
作者
He, Wentao [1 ,2 ]
Ren, Jianfeng [2 ,3 ]
Bai, Ruibin [2 ,3 ]
Jiang, Xudong [4 ]
机构
[1] Ningbo Univ, Fac Elect Engn & Comp Sci, Ningbo 315211, Peoples R China
[2] Univ Nottingham Ningbo China, Sch Comp Sci, Digital Port Technol Lab, 199 Taikang East Rd, Ningbo 315100, Peoples R China
[3] Univ Nottingham Ningbo China, Nottingham Ningbo China Beacons Excellence Res & I, 199 Taikang East Rd, Ningbo 315100, Peoples R China
[4] Nanyang Technol Univ, Sch Elect & Elect Engn, 50 Nanyang Ave, Singapore City 639798, Singapore
基金
中国国家自然科学基金;
关键词
Micro-Doppler signature; Radar gait recognition; Spectrogram; Cadence velocity diagram; Asymmetric Attention Fusion; IDENTIFICATION; NETWORKS; IMAGE;
D O I
10.1016/j.patcog.2024.111101
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video-based gait recognition suffers from potential privacy issues and performance degradation due to dim environments, partial occlusions, or camera view changes. Radar has recently become increasingly popular and overcome various challenges presented by vision sensors. To capture tiny differences in radar gait signatures of different people, a dual-branch Swin Transformer is proposed, where one branch captures the time variations of the radar micro-Doppler signature and the other captures the repetitive frequency patterns in the spectrogram. Unlike natural images where objects can be translated, rotated, or scaled, the spatial coordinates of spectrograms and CVDs have unique physical meanings, and there is no affine transformation for radar targets in these synthetic images. The patch splitting mechanism in Vision Transformer makes it ideal to extract discriminant information from patches, and learn the attentive information across patches, as each patch carries some unique physical properties of radar targets. Swin Transformer consists of a set of cascaded Swin blocks to extract semantic features from shallow to deep representations, further improving the classification performance. Lastly, to highlight the branch with larger discriminant power, an Asymmetric Attention Fusion is proposed to optimally fuse the discriminant features from the two branches. To enrich the research on radar gait recognition, a large-scale NTU-RGR dataset is constructed, containing 45,768 radar frames of 98 subjects. The proposed method is evaluated on the NTU-RGR dataset and the MMRGait-1.0 database. It consistently and significantly outperforms all the compared methods on both datasets. The codes are available at: https://github.com/wentaoheunnc/NTU-RGR.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] FLSTrack: focused linear attention swin-transformer network with dual-branch decoder for end-to-end multi-object tracking
    Zu, Dafu
    Duan, Xun
    Kong, Guangqian
    Long, Huiyun
    SIGNAL IMAGE AND VIDEO PROCESSING, 2025, 19 (01)
  • [22] Dual-Branch Multimodal Fusion Network for Driver Facial Emotion Recognition
    Wang, Le
    Chang, Yuchen
    Wang, Kaiping
    APPLIED SCIENCES-BASEL, 2024, 14 (20):
  • [23] Dual-ATME: Dual-Branch Attention Network for Micro-Expression Recognition
    Zhou, Haoliang
    Huang, Shucheng
    Li, Jingting
    Wang, Su-Jing
    ENTROPY, 2023, 25 (03)
  • [24] Learning upper patch attention using dual-branch training strategy for masked face recognition
    Zhang, Yuxuan
    Wang, Xin
    Shakeel, M. Saad
    Wan, Hao
    Kang, Wenxiong
    PATTERN RECOGNITION, 2022, 126
  • [25] A Multiscale Dual-Branch Feature Fusion and Attention Network for Hyperspectral Images Classification
    Gao, Hongmin
    Zhang, Yiyan
    Chen, Zhonghao
    Li, Chenming
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2021, 14 : 8180 - 8192
  • [26] TGDAUNet: Transformer and GCNN based dual-branch attention UNet for medical image segmentation
    Song, Pengfei
    Li, Jinjiang
    Fan, Hui
    Fan, Linwei
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 167
  • [27] Dual-branch feature fusion dehazing network via multispectral channel attention
    Jian, Huachun
    Zhang, Yongjun
    Gao, Weihao
    Wang, Bufan
    Wang, Guomei
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (07) : 2655 - 2671
  • [28] MDAN: Multilevel dual-branch attention network for infrared and visible image fusion
    Wang, Jiawei
    Jiang, Min
    Kong, Jun
    OPTICS AND LASERS IN ENGINEERING, 2024, 176
  • [29] A Dual-Branch Spatio-Temporal-Spectral Transformer Feature Fusion Network for EEG-based Visual Recognition
    Luo, Jie
    Cui, Weigang
    Xu, Song
    Wang, Lina
    Li, Xiao
    Liao, Xiaofeng
    Li, Yang
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (02) : 1721 - 1731
  • [30] DFTI: Dual-Branch Fusion Network Based on Transformer and Inception for Space Noncooperative Objects
    Zhang, Zhao
    Zhou, Dong
    Sun, Guanghui
    Hu, YuHui
    Deng, Runran
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73