CamRaDepth: Semantic Guided Depth Estimation Using Monocular Camera and Sparse Radar for Automotive Perception

被引:1
|
作者
Sauerbeck, Florian [1 ]
Halperin, Dan [1 ]
Connert, Lukas [1 ]
Betz, Johannes [2 ]
机构
[1] Tech Univ Munich, Munich Inst Robot & Machine Intelligence MIRMI, Inst Automot Technol, TUM Sch Engn & Design,Dept Mobil Syst Engn, D-85748 Garching, Germany
[2] Tech Univ Munich, Munich Inst Robot & Machine Intelligence MIRMI, TUM Sch Engn & Design, Dept Mobil Syst Engn,Professorship Autonomous Veh, D-85748 Garching, Germany
关键词
Autonomous driving; computer vision; depth prediction; intelligent vehicles; semantic segmentation; sensor fusion;
D O I
10.1109/JSEN.2023.3321886
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Our research aims to generate robust, dense 3-D depth maps for robotics, especially autonomous driving applications. Since cameras output 2-D images and active sensors such as LiDAR or radar produce sparse depth measurements, dense depth maps need to be estimated. Recent methods based on visual transformer networks have outperformed conventional deep learning approaches in various computer vision tasks, including depth prediction, but have focused on the use of a single camera image. This article explores the potential of visual transformers applied to the fusion of monocular images, semantic segmentation, and projected sparse radar reflections for robust monocular depth estimation. The addition of a semantic segmentation branch is used to add object-level understanding and is investigated in a supervised and unsupervised manner. We evaluate our new depth estimation approach on the nuScenes dataset where it outperforms existing state-of-the-art camera-radar depth estimation methods. We show that models can benefit from an additional segmentation branch during the training process by transfer learning even without running segmentation at inference. Further studies are needed to investigate the usage of 4-D-imaging radars and enhanced ground-truth generation in more detail. The related code is available as open-source software under https://github.com/TUMFTM/CamRaDepth.
引用
收藏
页码:28442 / 28453
页数:12
相关论文
共 50 条
  • [1] Expanding Sparse Radar Depth Based on Joint Bilateral Filter for Radar-Guided Monocular Depth Estimation
    Lo, Chen-Chou
    Vandewalle, Patrick
    [J]. SENSORS, 2024, 24 (06)
  • [2] Monocular Fisheye Camera Depth Estimation Using Sparse LiDAR Supervision
    Kumar, Varun Ravi
    Milz, Stefan
    Witt, Christian
    Simon, Martin
    Amende, Karl
    Petzold, Johannes
    Yogamani, Senthil
    Pech, Timo
    [J]. 2018 21ST INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2018, : 2853 - 2858
  • [3] Depth Estimation from Monocular Images and Sparse Radar Data
    Lin, Juan-Ting
    Dai, Dengxin
    Van Gool, Luc
    [J]. 2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 10233 - 10240
  • [4] Sparse depth densification for monocular depth estimation
    Zhen Liang
    Tiyu Fang
    Yanzhu Hu
    Yingjian Wang
    [J]. Multimedia Tools and Applications, 2024, 83 : 14821 - 14838
  • [5] DEPTH ESTIMATION FROM MONOCULAR IMAGES AND SPARSE RADAR USING DEEP ORDINAL REGRESSION NETWORK
    Lo, Chen-Chou
    Vandewalle, Patrick
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 3343 - 3347
  • [6] Sparse depth densification for monocular depth estimation
    Liang, Zhen
    Fang, Tiyu
    Hu, Yanzhu
    Wang, Yingjian
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (05) : 14821 - 14838
  • [7] Wearable Depth Camera: Monocular Depth Estimation via Sparse Optimization Under Weak Supervision
    He, Li
    Chen, Chuangbin
    Zhang, Tao
    Zhu, Haifei
    Wan, Shaohua
    [J]. IEEE ACCESS, 2018, 6 : 41337 - 41345
  • [8] Dynamic Semantically Guided Monocular Depth Estimation for UAV Environment Perception
    Miclea, Vlad-Cristian
    Nedevschi, Sergiu
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 11
  • [9] RCDformer: Transformer-based dense depth estimation by sparse radar and camera
    Huang, Xinyue
    Ma, Yongtao
    Yu, Zedong
    Zhao, Haibo
    [J]. NEUROCOMPUTING, 2024, 589
  • [10] Depth Estimation in Still Images and Videos Using a Motionless Monocular Camera
    Diamantas, Sotirios
    Astaras, Stefanos
    Pnevmatikakis, Aristodemos
    [J]. 2016 IEEE INTERNATIONAL CONFERENCE ON IMAGING SYSTEMS AND TECHNIQUES (IST), 2016, : 129 - 134