RadarFormer: End-to-End Human Perception With Through-Wall Radar and Transformers

被引:6
|
作者
Zheng, Zhijie [1 ,2 ]
Zhang, Diankun [1 ,2 ]
Liang, Xiao [1 ,2 ]
Liu, Xiaojun [1 ,2 ]
Fang, Guangyou [1 ,2 ]
机构
[1] Chinese Acad Sci, Aerosp Informat Res Inst, Key Lab Electromagnet Radiat & Detect Technol, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100039, Peoples R China
基金
中国国家自然科学基金;
关键词
End-to-end signal processing; fine-grained human perception; radio frequency (RF) signal; self-attention (SA) mechanism; ACTIVITY RECOGNITION; NETWORK;
D O I
10.1109/TNNLS.2023.3314031
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For fine-grained human perception tasks such as pose estimation and activity recognition, radar-based sensors show advantages over optical cameras in low-visibility, privacy-aware, and wall-occlusive environments. Radar transmits radio frequency signals to irradiate the target of interest and store the target information in the echo signals. One common approach is to transform the echoes into radar images and extract the features with convolutional neural networks. This article introduces RadarFormer, the first method that introduces the self-attention (SA) mechanism to perform human perception tasks directly from radar echoes. It bypasses the imaging algorithm and realizes end-to-end signal processing. Specifically, we give constructive proof that processing radar echoes using the SA mechanism is at least as expressive as processing radar images using the convolutional layer. On this foundation, we design RadarFormer, which is a Transformer-like model to process radar signals. It benefits from the fast-/slow-time SA mechanism considering the physical characteristics of radar signals. RadarFormer extracts human representations from radar echoes and handles various downstream human perception tasks. The experimental results demonstrate that our method outperforms the state-of-the-art radar-based methods both in performance and computational cost and obtains accurate human perception results even in dark and occlusive environments.
引用
收藏
页码:1 / 15
页数:15
相关论文
共 50 条
  • [21] On the Use of Transformers for End-to-End Optical Music Recognition
    Rios-Vila, Antonio
    Inesta, Jose M.
    Calvo-Zaragoza, Jorge
    PATTERN RECOGNITION AND IMAGE ANALYSIS (IBPRIA 2022), 2022, 13256 : 470 - 481
  • [22] VPDETR: End-to-End Vanishing Point DEtection TRansformers
    Chen, Taiyan
    Ying, Xianghua
    Yang, Jinfa
    Wang, Ruibin
    Guo, Ruohao
    Xing, Bowei
    Shi, Ji
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 2, 2024, : 1192 - 1200
  • [23] The End-to-End Segmentation on Automotive Radar Imagery
    Xiao, Yang
    Daniel, Liam
    Gashinova, Marina
    2021 18TH EUROPEAN RADAR CONFERENCE (EURAD), 2021, : 265 - 268
  • [24] Analysis of human backscattering in buildings for through-wall radar applications
    Thiel, Michael
    Sarabandi, Kamal
    2009 IEEE ANTENNAS AND PROPAGATION SOCIETY INTERNATIONAL SYMPOSIUM AND USNC/URSI NATIONAL RADIO SCIENCE MEETING, VOLS 1-6, 2009, : 3215 - 3218
  • [25] Using an End-to-End Convolutional Network on Radar Signal for Human Activity Classification
    Ye, Wenbin
    Chen, Haiquan
    Li, Bing
    IEEE SENSORS JOURNAL, 2019, 19 (24) : 12244 - 12252
  • [26] An End-to-End Network for Continuous Human Motion Recognition via Radar Radios
    Zhao, Running
    Ma, Xiaolin
    Liu, Xinhua
    Liu, Jian
    IEEE SENSORS JOURNAL, 2021, 21 (05) : 6487 - 6496
  • [27] Chaos Through-Wall Imaging Radar
    Xu H.
    Wang B.
    Zhang J.
    Liu L.
    Li Y.
    Wang Y.
    Wang A.
    Sensing and Imaging, 2017, 18 (1):
  • [28] Human similarity behaviour classification based on through-wall radar
    Huang, Ling
    Zeng, Hao
    Ma, Chegnyuan
    PROCEEDINGS OF 2023 7TH INTERNATIONAL CONFERENCE ON ELECTRONIC INFORMATION TECHNOLOGY AND COMPUTER ENGINEERING, EITCE 2023, 2023, : 806 - 811
  • [29] End-to-End Referring Video Object Segmentation with Multimodal Transformers
    Botach, Adam
    Zheltonozhskii, Evgenii
    Baskin, Chaim
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 4975 - 4985
  • [30] SWINBERT: End-to-End Transformers with Sparse Attention for Video Captioning
    Lin, Kevin
    Li, Linjie
    Lin, Chung-Ching
    Ahmed, Faisal
    Gan, Zhe
    Liu, Zicheng
    Lu, Yumao
    Wang, Lijuan
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 17928 - 17937