RadarFormer: End-to-End Human Perception With Through-Wall Radar and Transformers

被引:6
|
作者
Zheng, Zhijie [1 ,2 ]
Zhang, Diankun [1 ,2 ]
Liang, Xiao [1 ,2 ]
Liu, Xiaojun [1 ,2 ]
Fang, Guangyou [1 ,2 ]
机构
[1] Chinese Acad Sci, Aerosp Informat Res Inst, Key Lab Electromagnet Radiat & Detect Technol, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100039, Peoples R China
基金
中国国家自然科学基金;
关键词
End-to-end signal processing; fine-grained human perception; radio frequency (RF) signal; self-attention (SA) mechanism; ACTIVITY RECOGNITION; NETWORK;
D O I
10.1109/TNNLS.2023.3314031
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For fine-grained human perception tasks such as pose estimation and activity recognition, radar-based sensors show advantages over optical cameras in low-visibility, privacy-aware, and wall-occlusive environments. Radar transmits radio frequency signals to irradiate the target of interest and store the target information in the echo signals. One common approach is to transform the echoes into radar images and extract the features with convolutional neural networks. This article introduces RadarFormer, the first method that introduces the self-attention (SA) mechanism to perform human perception tasks directly from radar echoes. It bypasses the imaging algorithm and realizes end-to-end signal processing. Specifically, we give constructive proof that processing radar echoes using the SA mechanism is at least as expressive as processing radar images using the convolutional layer. On this foundation, we design RadarFormer, which is a Transformer-like model to process radar signals. It benefits from the fast-/slow-time SA mechanism considering the physical characteristics of radar signals. RadarFormer extracts human representations from radar echoes and handles various downstream human perception tasks. The experimental results demonstrate that our method outperforms the state-of-the-art radar-based methods both in performance and computational cost and obtains accurate human perception results even in dark and occlusive environments.
引用
收藏
页码:1 / 15
页数:15
相关论文
共 50 条
  • [41] PSTR: End-to-End One-Step Person Search With Transformers
    Cao, Jiale
    Pang, Yanwei
    Anwer, Rao Muhammad
    Cholakkal, Hisham
    Xie, Jin
    Shah, Mubarak
    Khan, Fahad Shahbaz
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9448 - 9457
  • [42] An Empirical Study of Training End-to-End Vision-and-Language Transformers
    Dou, Zi-Yi
    Xu, Yichong
    Gan, Zhe
    Wang, Jianfeng
    Wang, Shuohang
    Wang, Lijuan
    Zhu, Chenguang
    Zhang, Pengchuan
    Yuan, Lu
    Peng, Nanyun
    Liu, Zicheng
    Zeng, Michael
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 18145 - 18155
  • [43] End-to-End Ugly Duckling Sign Detection for Melanoma Identification with Transformers
    Yu, Zhen
    Mar, Victoria
    Eriksson, Anders
    Chandra, Shakes
    Bonnington, Paul
    Zhang, Lei
    Ge, Zongyuan
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT VII, 2021, 12907 : 176 - 184
  • [44] CurT: End-to-End Text Line Detection in Historical Documents with Transformers
    Kiessling, Benjamin
    FRONTIERS IN HANDWRITING RECOGNITION, ICFHR 2022, 2022, 13639 : 34 - 48
  • [45] Human Target Tracking for Small Aperture Through-wall Imaging Radar
    Li, Huquan
    Cui, Guolong
    Kong, Lingjiang
    Guo, Shisheng
    Wang, Mingyang
    Yang, Haining
    2019 IEEE RADAR CONFERENCE (RADARCONF), 2019,
  • [46] Through-Wall Detection of Human Being's Movement by UWB Radar
    Li, Jing
    Zeng, Zhaofa
    Sun, Jiguang
    Liu, Fengshan
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2012, 9 (06) : 1079 - 1083
  • [47] DwinFormer: Dual Window Transformers for End-to-End Monocular Depth Estimation
    Rahman, Md Awsafur
    Fattah, Shaikh Anowarul
    IEEE SENSORS JOURNAL, 2023, 23 (18) : 21443 - 21451
  • [48] End-to-End Video Snapshot Compressive Imaging using Video Transformers
    Saideni, Wael
    Courreges, Fabien
    Helbert, David
    Cances, Jean Pierre
    2022 ELEVENTH INTERNATIONAL CONFERENCE ON IMAGE PROCESSING THEORY, TOOLS AND APPLICATIONS (IPTA), 2022,
  • [49] MaX-DeepLab: End-to-End Panoptic Segmentation with Mask Transformers
    Wang, Huiyu
    Zhu, Yukun
    Adam, Hartwig
    Yuille, Alan
    Chen, Liang-Chieh
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 5459 - 5470
  • [50] End-to-End Spatio-Temporal Action Localisation with Video Transformers
    Gritsenko, Alexey A.
    Xiong, Xuehan
    Djolonga, Josip
    Dehghani, Mostafa
    Sun, Chen
    Lucic, Mario
    Schmid, Cordelia
    Arnab, Anurag
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 18373 - 18383