Depth-Enhanced Deep Learning Approach For Monocular Camera Based 3D Object Detection

被引:0
|
作者
Wang, Chuyao [1 ]
Aouf, Nabil [1 ]
机构
[1] City Univ London, Sch Sci & Technol, Dept Engn, London, England
关键词
3D Object detection; Autonomous driving; Machine learning;
D O I
10.1007/s10846-024-02128-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Automatic 3D object detection using monocular cameras presents significant challenges in the context of autonomous driving. Precise labeling of 3D object scales requires accurate spatial information, which is difficult to obtain from a single image due to the inherent lack of depth information in monocular images, compared to LiDAR data. In this paper, we propose a novel approach to address this issue by enhancing deep neural networks with depth information for monocular 3D object detection. The proposed method comprises three key components: 1)Feature Enhancement Pyramid Module: We extend the conventional Feature Pyramid Networks (FPN) by introducing a feature enhancement pyramid network. This module fuses feature maps from the original pyramid and captures contextual correlations across multiple scales. To increase the connectivity between low-level and high-level features, additional pathways are incorporated. 2)Auxiliary Dense Depth Estimator: We introduce an auxiliary dense depth estimator that generates dense depth maps to enhance the spatial perception capabilities of the deep network model without adding computational burden. 3)Augmented Center Depth Regression: To aid center depth estimation, we employ additional bounding box vertex depth regression based on geometry. Our experimental results demonstrate the superiority of the proposed technique over existing competitive methods reported in the literature. The approach showcases remarkable performance improvements in monocular 3D object detection, making it a promising solution for autonomous driving applications.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Absolute Distance Prediction Based on Deep Learning Object Detection and Monocular Depth Estimation Models
    Masoumian, Armin
    Marei, David G. F.
    Abdulwahab, Saddam
    Cristiano, Julian
    Puig, Domenec
    Rashwan, Hatem A.
    ARTIFICIAL INTELLIGENCE RESEARCH AND DEVELOPMENT, 2021, 339 : 325 - 334
  • [42] Monocular 3D Object Detection Based on Uncertainty Prediction of Keypoints
    Chen, Mu
    Zhao, Huaici
    Liu, Pengfei
    MACHINES, 2022, 10 (01)
  • [43] MonoDCN: Monocular 3D object detection based on dynamic convolution
    Qu, Shenming
    Yang, Xinyu
    Gao, Yiming
    Liang, Shengbin
    PLOS ONE, 2022, 17 (10):
  • [44] eGAC3D: enhancing depth adaptive convolution and depth estimation for monocular 3D object pose detection
    Ngo, Duc Tuan
    Bui, Minh-Quan Viet
    Nguyen, Duc Dung
    Pham, Hoang-Anh
    PEERJ COMPUTER SCIENCE, 2022, 8
  • [45] Triangulation Learning Network: from Monocular to Stereo 3D Object Detection
    Qin, Zengyi
    Wang, Jinglu
    Lu, Yan
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 7607 - 7615
  • [46] Monocular 3D Object Detection Utilizing Auxiliary Learning With Deformable Convolution
    Chen, Jiun-Han
    Shieh, Jeng-Lun
    Haq, Muhamad Amirul
    Ruan, Shanq-Jang
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (03) : 2424 - 2436
  • [47] A 3D Grid Mapping System Based on Depth Prediction from a Monocular Camera
    Yan, Peifeng
    Lan, Yuqing
    Yang, Shaowu
    2020 12TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTATIONAL INTELLIGENCE (ICACI), 2020, : 564 - 570
  • [48] Automotive Radar Dataset for Deep Learning Based 3D Object Detection
    Meyer, Michael
    Kuschk, Georg
    2019 16TH EUROPEAN RADAR CONFERENCE (EURAD), 2019, : 129 - 132
  • [49] 3D Object Detection from Point Cloud Based on Deep Learning
    Hao, Ning
    WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2022, 2022
  • [50] Rational 3D object placement based on deep learning based plane detection
    Yang, Wei-Jong
    Lu, Li-Yu
    Chan, Din-Yuen
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (28) : 44555 - 44576