Self-Supervised Deep Monocular Depth Estimation With Ambiguity Boosting

被引:9
|
作者
Bello, Juan Luis Gonzalez [1 ]
Kim, Munchurl [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Sch Elect Engn, Daejeon 34141, South Korea
关键词
Training; Boosting; Cameras; Estimation; Kernel; Task analysis; Image reconstruction; Monocular depth estimation; self-supervised learning; ambiguity boosting; STEREO;
D O I
10.1109/TPAMI.2021.3124079
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a novel two-stage training strategy with ambiguity boosting for the self-supervised learning of single view depths from stereo images. Our proposed two-stage learning strategy first aims to obtain a coarse depth prior by training an auto-encoder network for a stereoscopic view synthesis task. This prior knowledge is then boosted and used to self-supervise the model in the second stage of training in our novel ambiguity boosting loss. Our ambiguity boosting loss is a confidence-guided type of data augmentation loss that improves the accuracy and consistency of generated depth maps under several transformations of the single-image input. To show the benefits of the proposed two-stage training strategy with boosting, our two previous depth estimation (DE) networks, one with t-shaped adaptive kernels and the other with exponential disparity volumes, are extended with our new learning strategy, referred to as DBoosterNet-t and DBoosterNet-e, respectively. Our self-supervised DBoosterNets are competitive, and in some cases even better, compared to the most recent supervised SOTA methods, and are remarkably superior to the previous self-supervised methods for monocular DE on the challenging KITTI dataset. We present intensive experimental results, showing the efficacy of our method for the self-supervised monocular DE task.
引用
收藏
页码:9131 / 9149
页数:19
相关论文
共 50 条
  • [1] Deep Digging into the Generalization of Self-Supervised Monocular Depth Estimation
    Bae, Jinwoo
    Moon, Sungho
    Im, Sunghoon
    [J]. THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 1, 2023, : 187 - 196
  • [2] Digging Into Self-Supervised Monocular Depth Estimation
    Godard, Clement
    Mac Aodha, Oisin
    Firman, Michael
    Brostow, Gabriel
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 3827 - 3837
  • [3] On the uncertainty of self-supervised monocular depth estimation
    Poggi, Matteo
    Aleotti, Filippo
    Tosi, Fabio
    Mattoccia, Stefano
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 3224 - 3234
  • [4] Revisiting Self-supervised Monocular Depth Estimation
    Kim, Ue-Hwan
    Lee, Gyeong-Min
    Kim, Jong-Hwan
    [J]. ROBOT INTELLIGENCE TECHNOLOGY AND APPLICATIONS 6, 2022, 429 : 336 - 350
  • [5] Self-supervised monocular depth estimation in fog
    Tao, Bo
    Hu, Jiaxin
    Jiang, Du
    Li, Gongfa
    Chen, Baojia
    Qian, Xinbo
    [J]. OPTICAL ENGINEERING, 2023, 62 (03)
  • [6] Semantically guided self-supervised monocular depth estimation
    Lu, Xiao
    Sun, Haoran
    Wang, Xiuling
    Zhang, Zhiguo
    Wang, Haixia
    [J]. IET IMAGE PROCESSING, 2022, 16 (05) : 1293 - 1304
  • [7] Self-supervised deep monocular visual odometry and depth estimation with observation variation
    Zhao, Wentao
    Wang, Yanbo
    Wang, Zehao
    Li, Rui
    Xiao, Peng
    Wang, Jingchuan
    Guo, Rui
    [J]. DISPLAYS, 2023, 80
  • [8] Self-supervised monocular depth estimation in dynamic scenes based on deep learning
    Cheng, Binbin
    Yu, Ying
    Zhang, Lei
    Wang, Ziquan
    Jiang, Zhipeng
    [J]. National Remote Sensing Bulletin, 2024, 28 (09) : 2170 - 2186
  • [9] Joint Self-Supervised Monocular Depth Estimation and SLAM
    Xing, Xiaoxia
    Cai, Yinghao
    Lu, Tao
    Yang, Yiping
    Wen, Dayong
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 4030 - 4036
  • [10] Self-Supervised Monocular Scene Decomposition and Depth Estimation
    Safadoust, Sadra
    Guney, Fatma
    [J]. 2021 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2021), 2021, : 627 - 636