RadarFormer: End-to-End Human Perception With Through-Wall Radar and Transformers

被引:6
|
作者
Zheng, Zhijie [1 ,2 ]
Zhang, Diankun [1 ,2 ]
Liang, Xiao [1 ,2 ]
Liu, Xiaojun [1 ,2 ]
Fang, Guangyou [1 ,2 ]
机构
[1] Chinese Acad Sci, Aerosp Informat Res Inst, Key Lab Electromagnet Radiat & Detect Technol, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100039, Peoples R China
基金
中国国家自然科学基金;
关键词
End-to-end signal processing; fine-grained human perception; radio frequency (RF) signal; self-attention (SA) mechanism; ACTIVITY RECOGNITION; NETWORK;
D O I
10.1109/TNNLS.2023.3314031
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For fine-grained human perception tasks such as pose estimation and activity recognition, radar-based sensors show advantages over optical cameras in low-visibility, privacy-aware, and wall-occlusive environments. Radar transmits radio frequency signals to irradiate the target of interest and store the target information in the echo signals. One common approach is to transform the echoes into radar images and extract the features with convolutional neural networks. This article introduces RadarFormer, the first method that introduces the self-attention (SA) mechanism to perform human perception tasks directly from radar echoes. It bypasses the imaging algorithm and realizes end-to-end signal processing. Specifically, we give constructive proof that processing radar echoes using the SA mechanism is at least as expressive as processing radar images using the convolutional layer. On this foundation, we design RadarFormer, which is a Transformer-like model to process radar signals. It benefits from the fast-/slow-time SA mechanism considering the physical characteristics of radar signals. RadarFormer extracts human representations from radar echoes and handles various downstream human perception tasks. The experimental results demonstrate that our method outperforms the state-of-the-art radar-based methods both in performance and computational cost and obtains accurate human perception results even in dark and occlusive environments.
引用
收藏
页码:1 / 15
页数:15
相关论文
共 50 条
  • [1] End-to-End Human Pose and Mesh Reconstruction with Transformers
    Lin, Kevin
    Wang, Lijuan
    Liu, Zicheng
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 1954 - 1963
  • [2] End-to-End Human-Gaze-Target Detection with Transformers
    Tu, Danyang
    Min, Xiongkuo
    Duan, Huiyu
    Guo, Guodong
    Zhai, Guangtao
    Shen, Wei
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 2192 - 2200
  • [3] End-to-end Symbolic Regression with Transformers
    Kamienny, Pierre-Alexandre
    d'Ascoli, Stephane
    Lample, Guillaume
    Charton, Francois
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [4] CPFormer: End-to-End Multi-Person Human Pose Estimation From Raw Radar Cubes With Transformers
    Chen, Lin
    Wang, Guoli
    IEEE SENSORS JOURNAL, 2025, 25 (07) : 12466 - 12478
  • [5] HOTR: End-to-End Human-Object Interaction Detection with Transformers
    Kim, Bumsoo
    Lee, Junhyun
    Kang, Jaewoo
    Kim, Eun-Sol
    Kim, Hyunwoo J.
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 74 - 83
  • [6] TransVG: End-to-End Visual Grounding with Transformers
    Deng, Jiajun
    Yang, Zhengyuan
    Chen, Tianlang
    Zhou, Wengang
    Li, Houqiang
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 1749 - 1759
  • [7] SYNCHRONOUS TRANSFORMERS FOR END-TO-END SPEECH RECOGNITION
    Tian, Zhengkun
    Yi, Jiangyan
    Bai, Ye
    Tao, Jianhua
    Zhang, Shuai
    Wen, Zhengqi
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7884 - 7888
  • [8] End-to-end Lane Shape Prediction with Transformers
    Liu, Ruijin
    Yuan, Zejian
    Liu, Tie
    Xiong, Zhiliang
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, : 3693 - 3701
  • [9] End-to-End Video Instance Segmentation with Transformers
    Wang, Yuqing
    Xu, Zhaoliang
    Wang, Xinlong
    Shen, Chunhua
    Cheng, Baoshan
    Shen, Hao
    Xia, Huaxia
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 8737 - 8746
  • [10] Cascade Transformers for End-to-End Person Search
    Yu, Rui
    Du, Dawei
    LaLonde, Rodney
    Davila, Daniel
    Funk, Christopher
    Hoogs, Anthony
    Clipp, Brian
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 7257 - 7266