Point Cloud 3D Object Detection Based on Improved SECOND Algorithm

被引:0
|
作者
Zhang Ying [1 ]
Jiang Liangliang [1 ]
Zhang Dongbo [1 ]
Duan Wanlin [1 ]
Sun Yue [1 ]
机构
[1] Xiangtan Univ, Coll Automat & Elect Informat, Xiangtan 411105, Hunan, Peoples R China
关键词
autonomous driving; three-dimensional object detection; feature fusion; loss function;
D O I
10.3788/LOP231016
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Rapid identification and precise positioning of surrounding targets are prerequisites and represent the foundation for safe autonomous vehicle driving. A point cloud 3D object detection algorithm based on an improved SECOND algorithm is proposed to address the challenges of inaccurate recognition and positioning in voxel-based point cloud 3D object detection methods. First, an adaptive spatial feature fusion module is introduced into a 2D convolutional backbone network to fuse spatial features of different scales, so as to improve the model's feature expression capability. Second, by fully utilizing the correlation between bounding box parameters, the three-dimensional distance-intersection over union (3D DIoU) is adopted as the bounding box localization regression loss function, thus improving regression task efficiency. Finally, considering both the classification confidence and positioning accuracy of candidate boxes, a new candidate box quality evaluation standard is utilized to obtain smoother regression results. Experimental results on the KITTI test set demonstrate that the 3D detection accuracy of the proposed algorithm is superior to many previous algorithms. Compared with the SECOND benchmark algorithm, the car and cyclist classes improves by 2. 86 and 3. 84 percentage points, respectively, under simple difficulty; 2. 99 and 3. 89 percentage points, respectively, under medium difficulty; and 7. 06 and 4. 27 percentage points, respectively, under difficult difficulty.
引用
收藏
页数:10
相关论文
共 24 条
  • [1] Lidar 3D Target Detection Based on Improved PointPillars
    Chen Dejiang
    Yu Wenjun
    Gao Yongbin
    [J]. LASER & OPTOELECTRONICS PROGRESS, 2023, 60 (10)
  • [2] Geiger A, 2012, PROC CVPR IEEE, P3354, DOI 10.1109/CVPR.2012.6248074
  • [3] Fast R-CNN
    Girshick, Ross
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1440 - 1448
  • [4] DVFENet: Dual-branch voxel feature extraction network for 3D object detection
    He, Yunqian
    Xia, Guihua
    Luo, Yongkang
    Su, Li
    Zhang, Zhi
    Li, Wanyi
    Wang, Peng
    [J]. NEUROCOMPUTING, 2021, 459 : 201 - 211
  • [5] Ku J, 2018, IEEE INT C INT ROBOT, P5750, DOI 10.1109/IROS.2018.8594049
  • [6] PointPillars: Fast Encoders for Object Detection from Point Clouds
    Lang, Alex H.
    Vora, Sourabh
    Caesar, Holger
    Zhou, Lubing
    Yang, Jiong
    Beijbom, Oscar
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 12689 - 12697
  • [7] 3D Object Detection Based on Improved Frustum PointNet
    Liu Xunhua
    Sun Shaoyuan
    Gu Lipeng
    Li Xiang
    [J]. LASER & OPTOELECTRONICS PROGRESS, 2020, 57 (20)
  • [8] AutoShape: Real-Time Shape-Aware Monocular 3D Object Detection
    Liu, Zongdai
    Zhou, Dingfu
    Lu, Feixiang
    Fang, Jin
    Zhang, Liangjun
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 15621 - 15630
  • [9] Frustum PointNets for 3D Object Detection from RGB-D Data
    Qi, Charles R.
    Liu, Wei
    Wu, Chenxia
    Su, Hao
    Guibas, Leonidas J.
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 918 - 927
  • [10] Qi CR, 2017, Arxiv, DOI arXiv:1706.02413