SMURF: Spatial Multi-Representation Fusion for 3D Object Detection With 4D Imaging Radar

被引:16
|
作者
Liu, Jianan [2 ]
Zhao, Qiuchi [1 ]
Xiong, Weiyi [1 ]
Huang, Tao [3 ]
Han, Qing-Long [4 ]
Zhu, Bing [1 ]
机构
[1] Beihang Univ, Sch Automat Sci & Elect Engn, Beijing 100191, Peoples R China
[2] Vitalent Consulting, S-41761 Gothenburg, Sweden
[3] James Cook Univ, Coll Sci & Engn, Cairns, Qld 4878, Australia
[4] Swinburne Univ Technol, Sch Sci Comp & Engn Technol, Melbourne, Vic 3122, Australia
来源
基金
澳大利亚研究理事会; 中国国家自然科学基金;
关键词
Radar; Radar imaging; Point cloud compression; Radar detection; Feature extraction; Three-dimensional displays; Object detection; 4D imaging radar; radar point cloud; kernel density estimation; multi-dimensional Gaussian mixture; 3D object detection; autonomous driving; MIMO RADAR; NETWORK; CNN;
D O I
10.1109/TIV.2023.3322729
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The 4D millimeter-Wave (mmWave) radar is a promising technology for vehicle sensing due to its cost-effectiveness and operability in adverse weather conditions. However, the adoption of this technology has been hindered by sparsity and noise issues in radar point cloud data. This article introduces spatial multi-representation fusion (SMURF), a novel approach to 3D object detection using a single 4D imaging radar. SMURF leverages multiple representations of radar detection points, including pillarization and density features of a multi-dimensional Gaussian mixture distribution through kernel density estimation (KDE). KDE effectively mitigates measurement inaccuracy caused by limited angular resolution and multi-path propagation of radar signals. Additionally, KDE helps alleviate point cloud sparsity by capturing density features. Experimental evaluations on View-of-Delft (VoD) and TJ4DRadSet datasets demonstrate the effectiveness and generalization ability of SMURF, outperforming recently proposed 4D imaging radar-based single-representation models. Moreover, while using 4D imaging radar only, SMURF still achieves comparable performance to the state-of-the-art 4D imaging radar and camera fusion-based method, with an increase of 1.22% in the mean average precision on bird's-eye view of TJ4DRadSet dataset and 1.32% in the 3D mean average precision on the entire annotated area of VoD dataset. Our proposed method demonstrates impressive inference time and addresses the challenges of real-time detection, with the inference time no more than 0.05 seconds for most scans on both datasets. This research highlights the benefits of 4D mmWave radar and is a strong benchmark for subsequent works regarding 3D object detection with 4D imaging radar. Index Terms-4D imaging radar, radar point cloud,
引用
收藏
页码:799 / 812
页数:14
相关论文
共 50 条
  • [31] CenterTransFuser: radar point cloud and visual information fusion for 3D object detection
    Yan Li
    Kai Zeng
    Tao Shen
    EURASIP Journal on Advances in Signal Processing, 2023
  • [32] CenterTransFuser: radar point cloud and visual information fusion for 3D object detection
    Li, Yan
    Zeng, Kai
    Shen, Tao
    EURASIP JOURNAL ON ADVANCES IN SIGNAL PROCESSING, 2023, 2023 (01)
  • [33] CenterFusion: Center-based Radar and Camera Fusion for 3D Object Detection
    Nabati, Ramin
    Qi, Hairong
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 1526 - 1535
  • [34] LiRaFusion: Deep Adaptive LiDAR-Radar Fusion for 3D Object Detection
    Song, Jingyu
    Zhao, Lingjun
    Skinner, Katherine A.
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2024), 2024, : 18250 - 18257
  • [35] Multi-Task Multi-Sensor Fusion for 3D Object Detection
    Liang, Ming
    Yang, Bin
    Chen, Yun
    Hu, Rui
    Urtasun, Raquel
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 7337 - 7345
  • [36] Camera-Radar Fusion with Modality Interaction and Radar Gaussian Expansion for 3D Object Detection
    Liu, Xiang
    Li, Zhenglin
    Zhou, Yang
    Peng, Yan
    Luo, Jun
    CYBORG AND BIONIC SYSTEMS, 2024, 5
  • [37] Synthesis-based imaging-differentiation representation learning for multi-sequence 3D/4D MRI
    Han, Luyi
    Tan, Tao
    Zhang, Tianyu
    Huang, Yunzhi
    Wang, Xin
    Gao, Yuan
    Teuwen, Jonas
    Mann, Ritse
    MEDICAL IMAGE ANALYSIS, 2024, 92
  • [38] MLF3D: Multi-Level Fusion for Multi-Modal 3D Object Detection
    Jiang, Han
    Wang, Jianbin
    Xiao, Jianru
    Zhao, Yanan
    Chen, Wanqing
    Ren, Yilong
    Yu, Haiyang
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 1588 - 1593
  • [39] 4D Contrastive Superflows are Dense 3D Representation Learners
    Xu, Xiang
    Kong, Lingdong
    Shuai, Hui
    Zhang, Wenwei
    Pan, Liang
    Chen, Kai
    Liu, Ziwei
    Liu, Qingshan
    COMPUTER VISION-ECCV 2024, PT I, 2025, 15059 : 58 - 80
  • [40] Physical modeling of 3D and 4D laser imaging
    Anna, Guillaume
    Hamoir, Dominique
    Hespel, Laurent
    Lafay, Fabien
    Riviere, Nicolas
    Tanguy, Bernard
    LASER RADAR TECHNOLOGY AND APPLICATIONS XV, 2010, 7684