Object detection algorithm based on image and point cloud fusion with N3D_DIOU

被引:0
|
作者
Guo B.-Q. [1 ,2 ]
Xie G.-F. [1 ]
机构
[1] School of Mechanical, Electronic and Control Engineering, Beijing Jiaotong University, Beijing
[2] Frontiers Science Center for Smart High-speed Railway System, Beijing Jiaotong University, Beijing
关键词
2D image; 3D point cloud; 3D_DIOU; Feature fusion; Object detection;
D O I
10.37188/OPE.20212911.2703
中图分类号
学科分类号
摘要
Object detection is the basis of autonomous driving and robot navigation. To solve the problems of insufficient information in 2D images and the large data volume, uneven density, and low detection accuracy of 3D point clouds, a new 3D object-detection network is proposed through an image and point-cloud fusion with deep learning. To reduce the calculation load, the original point cloud is first filtered with the flat interceptor corresponding to the object's frame detected in the 2D image. To address the uneven density, an improved voting model network, based on a generalized Hough transform, is proposed for multiscale feature extraction. Finally, Normal Three-Dimensional Distance Intersection over Union (N3D_DIOU), a novel loss function, is extended from the Two-Dimensional Distance Intersection over Union (2D DIOU) loss function, which improves the consistency between the generated and target frames, and also improves the object-detection accuracy of the point cloud. Experiments on the KITTI dataset show that our algorithm improves the accuracy of three-dimensional detection by 0.71%, and the aerial-view detection accuracy by 7.28%, over outstanding classical methods. © 2021, Science Press. All right reserved.
引用
收藏
页码:2703 / 2713
页数:10
相关论文
共 31 条
  • [1] CHEN X Z, MA H M, WAN J, Et al., Multi-view 3D object detection network for autonomous driving[C], 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2126, pp. 6526-6534, (2017)
  • [2] GUO B Q, YU Z J, ZHANG N, Et al., 3D point cloud segmentation, classification and recognition algorithm of railway scene, Chinese Journal of Scientific Instrument, 38, 9, pp. 2103-2111, (2017)
  • [3] HE K M, ZHANG X Y, REN S Q, Et al., Spatial pyramid pooling in deep convolutional networks for visual recognition, IEEE Transactions on Pattern Analysis and Machine Intelligence, 37, 9, pp. 1904-1916, (2015)
  • [4] GIRSHICK R, DONAHUE J, DARRELL T, Et al., Rich feature hierarchies for accurate object detection and semantic segmentation[C], 2014 IEEE Conference on Computer Vision and Pattern Recognition, 2328, pp. 580-587, (2014)
  • [5] GIRSHICK R., Fast R-CNN, 2015 IEEE International Conference on Computer Vision (ICCV), 713, pp. 1440-1448, (2015)
  • [6] REN S Q, HE K M, GIRSHICK R, Et al., Faster R-CNN: towards real-time object detection with region proposal networks, IEEE Transactions on Pattern Analysis and Machine Intelligence, 39, 6, pp. 1137-1149, (2017)
  • [7] REDMON J, DIVVALA S, GIRSHICK R, Et al., You only look once: unified, real-time object detection, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2730, pp. 779-788, (2016)
  • [8] GUO B Q, GENG G, ZHU L Q, Et al., High-speed railway intruding object image generating with generative adversarial networks, Sensors, 19, 14, (2019)
  • [9] WANG J L, FU X S, HUANG ZH CH, Et al., Multi-type cooperative targets detection using improved YOLOv2 convolutional neural network, Opt. Precision Eng, 28, 1, pp. 251-260, (2020)
  • [10] ARNOLD E, AL-JARRAH O Y, DIANATI M, Et al., A survey on 3D object detection methods for autonomous driving applications, IEEE Transactions on Intelligent Transportation Systems, 20, 10, pp. 3782-3795, (2019)