RadarFormer: End-to-End Human Perception With Through-Wall Radar and Transformers

被引:6
|
作者
Zheng, Zhijie [1 ,2 ]
Zhang, Diankun [1 ,2 ]
Liang, Xiao [1 ,2 ]
Liu, Xiaojun [1 ,2 ]
Fang, Guangyou [1 ,2 ]
机构
[1] Chinese Acad Sci, Aerosp Informat Res Inst, Key Lab Electromagnet Radiat & Detect Technol, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100039, Peoples R China
基金
中国国家自然科学基金;
关键词
End-to-end signal processing; fine-grained human perception; radio frequency (RF) signal; self-attention (SA) mechanism; ACTIVITY RECOGNITION; NETWORK;
D O I
10.1109/TNNLS.2023.3314031
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For fine-grained human perception tasks such as pose estimation and activity recognition, radar-based sensors show advantages over optical cameras in low-visibility, privacy-aware, and wall-occlusive environments. Radar transmits radio frequency signals to irradiate the target of interest and store the target information in the echo signals. One common approach is to transform the echoes into radar images and extract the features with convolutional neural networks. This article introduces RadarFormer, the first method that introduces the self-attention (SA) mechanism to perform human perception tasks directly from radar echoes. It bypasses the imaging algorithm and realizes end-to-end signal processing. Specifically, we give constructive proof that processing radar echoes using the SA mechanism is at least as expressive as processing radar images using the convolutional layer. On this foundation, we design RadarFormer, which is a Transformer-like model to process radar signals. It benefits from the fast-/slow-time SA mechanism considering the physical characteristics of radar signals. RadarFormer extracts human representations from radar echoes and handles various downstream human perception tasks. The experimental results demonstrate that our method outperforms the state-of-the-art radar-based methods both in performance and computational cost and obtains accurate human perception results even in dark and occlusive environments.
引用
收藏
页码:1 / 15
页数:15
相关论文
共 50 条
  • [31] VRDFormer: End-to-End Video Visual Relation Detection with Transformers
    Zheng, Sipeng
    Chen, Shizhe
    Jin, Qin
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 18814 - 18824
  • [32] End-to-End Multi-Person Pose Estimation with Transformers
    Shi, Dahu
    Wei, Xing
    Li, Liangqi
    Ren, Ye
    Tan, Wenming
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 11059 - 11068
  • [33] Deeply Tensor Compressed Transformers for End-to-End Object Detection
    Zhen, Peining
    Gao, Ziyang
    Hou, Tianshu
    Cheng, Yuan
    Chen, Hai-Bao
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 4716 - 4724
  • [34] SimHumalator: An Open-Source End-to-End Radar Simulator for Human Activity Recognition
    Vishwakarma, Shelly
    Li, Wenda
    Tang, Chong
    Woodbridge, Karl
    Adve, Raviraj
    Chetty, Kevin
    IEEE AEROSPACE AND ELECTRONIC SYSTEMS MAGAZINE, 2022, 37 (03) : 6 - 22
  • [35] Through-wall detection of human being with compressed UWB radar data
    Baoju Zhang
    Wei Wang
    EURASIP Journal on Wireless Communications and Networking, 2013
  • [36] End-to-End Thorough Body Perception for Person Search
    Tian, Kun
    Huang, Houjing
    Ye, Yun
    Li, Shiyu
    Lin, Jinbin
    Huang, Guan
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 12079 - 12086
  • [37] Through-wall detection of human being with compressed UWB radar data
    Zhang, Baoju
    Wang, Wei
    EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2013,
  • [38] End-to-end Contextual Perception and Prediction with Interaction Transformer
    Li, Lingyun Luke
    Yang, Bin
    Liang, Ming
    Zeng, Wenyuan
    Ren, Mengye
    Segal, Sean
    Urtasun, Raquel
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 5784 - 5791
  • [39] Through-wall human being detection using UWB impulse radar
    Liang, Xiaolin
    Lv, Tingting
    Zhang, Hao
    Gao, Yong
    Fang, Guangyou
    EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2018,
  • [40] An Empirical Study of Training End-to-End Vision-and-Language Transformers
    Dou, Zi-Yi
    Xu, Yichong
    Gan, Zhe
    Wang, Jianfeng
    Wang, Shuohang
    Wang, Lijuan
    Zhu, Chenguang
    Zhang, Pengchuan
    Yuan, Lu
    Peng, Nanyun
    Liu, Zicheng
    Zeng, Michael
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 18145 - 18155