3D Object Detection Based on Extremely Sparse Laser Point Cloud and RGB Images

被引:1
|
作者
Qin Chao [1 ,2 ]
Wang Yafei [1 ]
Zhang Yuchao [2 ]
Yin Chengliang [1 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Mech Engn, Shanghai 200240, Peoples R China
[2] Shanghai Intelligent & Connected Vehicle R&D Ctr, Shanghai 201499, Peoples R China
关键词
remote sensing; laser point cloud; convolutional neural network; key point detection; deep learning;
D O I
10.3788/LOP202259.1828004
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The task of detecting 3D objects in complex traffic scenes is crucial and challenging. To address the high-cost problem of high-definition LiDAR and the poor effect of detection algorithms based on the millimeter wave radar and cameras used in mainstream detection algorithms, this study proposes a 3D target detection algorithm using low-definition LiDAR and a camera, which can significantly reduce the hardware cost of autonomous driving. To obtain a depth map, the 64-line LiDAR point cloud is first downsampled to 10% of the original point clouds, resulting in an extremely sparse point cloud, and fed to the depth-completion network with RGB images. Then, a point cloud bird-eye view is generated from the depth map based on the proposed algorithm for calculating the point cloud intensity. Finally, the point cloud bird-eye view is fed into the detection network to obtain the geometric information, heading angle, and category of the target stereo hounding box. The different algorithms are experimentally validated using KITT1 dataset. The experimental results demonstrate that the proposed algorithm can outperform some conventional high-definition LiDAR-based detection algorithms in terms of detection accuracy.
引用
收藏
页数:12
相关论文
共 31 条
  • [1] Bochkovskiy A, 2020, Arxiv, DOI [arXiv:2004.10934, DOI 10.48550/ARXIV.2004.10934, 10.48550/arXiv.2004.10934]
  • [2] Multi-View 3D Object Detection Network for Autonomous Driving
    Chen, Xiaozhi
    Ma, Huimin
    Wan, Ji
    Li, Bo
    Xia, Tian
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 6526 - 6534
  • [3] Geiger A, 2012, PROC CVPR IEEE, P3354, DOI 10.1109/CVPR.2012.6248074
  • [4] Fast R-CNN
    Girshick, Ross
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1440 - 1448
  • [5] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [6] J”rgensen E, 2019, Arxiv, DOI arXiv:1906.08070
  • [7] Ku J, 2018, IEEE INT C INT ROBOT, P5750, DOI 10.1109/IROS.2018.8594049
  • [8] Law H, 2020, Arxiv, DOI arXiv:1904.08900
  • [9] CornerNet: Detecting Objects as Paired Keypoints
    Law, Hei
    Deng, Jia
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2020, 128 (03) : 642 - 656
  • [10] Li B, 2016, Arxiv, DOI arXiv:1608.07916