Radar gait recognition using Dual-branch Swin Transformer with Asymmetric Attention Fusion

被引:0
|
作者
He, Wentao [1 ,2 ]
Ren, Jianfeng [2 ,3 ]
Bai, Ruibin [2 ,3 ]
Jiang, Xudong [4 ]
机构
[1] Ningbo Univ, Fac Elect Engn & Comp Sci, Ningbo 315211, Peoples R China
[2] Univ Nottingham Ningbo China, Sch Comp Sci, Digital Port Technol Lab, 199 Taikang East Rd, Ningbo 315100, Peoples R China
[3] Univ Nottingham Ningbo China, Nottingham Ningbo China Beacons Excellence Res & I, 199 Taikang East Rd, Ningbo 315100, Peoples R China
[4] Nanyang Technol Univ, Sch Elect & Elect Engn, 50 Nanyang Ave, Singapore City 639798, Singapore
基金
中国国家自然科学基金;
关键词
Micro-Doppler signature; Radar gait recognition; Spectrogram; Cadence velocity diagram; Asymmetric Attention Fusion; IDENTIFICATION; NETWORKS; IMAGE;
D O I
10.1016/j.patcog.2024.111101
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video-based gait recognition suffers from potential privacy issues and performance degradation due to dim environments, partial occlusions, or camera view changes. Radar has recently become increasingly popular and overcome various challenges presented by vision sensors. To capture tiny differences in radar gait signatures of different people, a dual-branch Swin Transformer is proposed, where one branch captures the time variations of the radar micro-Doppler signature and the other captures the repetitive frequency patterns in the spectrogram. Unlike natural images where objects can be translated, rotated, or scaled, the spatial coordinates of spectrograms and CVDs have unique physical meanings, and there is no affine transformation for radar targets in these synthetic images. The patch splitting mechanism in Vision Transformer makes it ideal to extract discriminant information from patches, and learn the attentive information across patches, as each patch carries some unique physical properties of radar targets. Swin Transformer consists of a set of cascaded Swin blocks to extract semantic features from shallow to deep representations, further improving the classification performance. Lastly, to highlight the branch with larger discriminant power, an Asymmetric Attention Fusion is proposed to optimally fuse the discriminant features from the two branches. To enrich the research on radar gait recognition, a large-scale NTU-RGR dataset is constructed, containing 45,768 radar frames of 98 subjects. The proposed method is evaluated on the NTU-RGR dataset and the MMRGait-1.0 database. It consistently and significantly outperforms all the compared methods on both datasets. The codes are available at: https://github.com/wentaoheunnc/NTU-RGR.
引用
收藏
页数:14
相关论文
共 50 条
  • [11] ATTENTION-BASED DUAL-STREAM VISION TRANSFORMER FOR RADAR GAIT RECOGNITION
    Chen, Shiliang
    He, Wentao
    Ren, Jianfeng
    Jiang, Xudong
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3668 - 3672
  • [12] A Dual-Branch Model Integrating CNN and Swin Transformer for Efficient Apple Leaf Disease Classification
    Si, Haiping
    Li, Mingchun
    Li, Weixia
    Zhang, Guipei
    Wang, Ming
    Li, Feitao
    Li, Yanling
    AGRICULTURE-BASEL, 2024, 14 (01):
  • [13] DUAL-BRANCH ATTENTION-IN-ATTENTION TRANSFORMER FOR SINGLE-CHANNEL SPEECH ENHANCEMENT
    Yu, Guochen
    Li, Andong
    Zheng, Chengshi
    Guo, Yinuo
    Wang, Yutian
    Wang, Hui
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7847 - 7851
  • [14] A dual-branch and dual attention transformer and CNN hybrid network for ultrasound image segmentation
    Zhang, Chong
    Wang, Lingtong
    Wei, Guohui
    Kong, Zhiyong
    Qiu, Min
    FRONTIERS IN PHYSIOLOGY, 2024, 15
  • [15] Dual-branch collaborative transformer for effective
    Qi, Xuanyu
    Song, Tianyu
    Dong, Haobo
    Jin, Jiyu
    Jin, Guiyue
    Li, Pengpeng
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 100
  • [16] MDTrans: Multi-scale and dual-branch feature fusion network based on Swin Transformer for building extraction in remote sensing images
    Diao, Kuo
    Zhu, Jinlong
    Liu, Guangjie
    Li, Meng
    IET IMAGE PROCESSING, 2024, 18 (11) : 2930 - 2942
  • [17] Dual-branch adaptive attention transformer for occluded person re-identification
    Lu, Yunhua
    Jiang, Mingzi
    Liu, Zhi
    Mu, Xinyu
    IMAGE AND VISION COMPUTING, 2023, 131
  • [18] Dual-branch self-attention network for pedestrian attribute recognition
    Liu, Zhenyu
    Zhang, Zhang
    Li, Da
    Zhang, Peng
    Shan, Caifeng
    PATTERN RECOGNITION LETTERS, 2022, 163 : 112 - 120
  • [19] An attention-based RGBD dual-branch gesture recognition network
    Chen, Bo
    Xie, Pengwei
    Hao, Nan
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 8022 - 8027
  • [20] A Dual-Branch Dynamic Graph Convolution Based Adaptive TransFormer Feature Fusion Network for EEG Emotion Recognition
    Sun, Mingyi
    Cui, Weigang
    Yu, Shuyue
    Han, Hongbin
    Hu, Bin
    Li, Yang
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2022, 13 (04) : 2218 - 2228