Monocular Depth Estimation Using Laplacian Pyramid-Based Depth Residuals

被引:139
|
作者
Song, Minsoo [1 ]
Lim, Seokjae [1 ]
Kim, Wonjun [1 ]
机构
[1] Konkuk Univ, Dept Elect & Elect Engn, Seoul 05029, South Korea
关键词
Estimation; Laplace equations; Decoding; Feature extraction; Convolution; Color; Image reconstruction; Monocular depth estimation; depth residuals; depth boundary; Laplacian pyramid; weight standardization; IMAGE; MODEL;
D O I
10.1109/TCSVT.2021.3049869
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
With a great success of the generative model via deep neural networks, monocular depth estimation has been actively studied by exploiting various encoder-decoder architectures. However, the decoding process in most previous methods, which repeats simple up-sampling operations, probably fails to fully utilize underlying properties of well-encoded features for monocular depth estimation. To resolve this problem, we propose a simple but effective scheme by incorporating the Laplacian pyramid into the decoder architecture. Specifically, encoded features are fed into different streams for decoding depth residuals, which are defined by decomposition of the Laplacian pyramid, and corresponding outputs are progressively combined to reconstruct the final depth map from coarse to fine scales. This is fairly desirable to precisely estimate the depth boundary as well as the global layout. We also propose to apply weight standardization to pre-activation convolution blocks of the decoder architecture, which gives a great help to improve the flow of gradients and thus makes optimization easier. Experimental results on benchmark datasets constructed under various indoor and outdoor environments demonstrate that the proposed method is effective for monocular depth estimation compared to state-of-the-art models. The code and model are publicly available at: |https://github.com/tjqansthd/LapDepth-release|.
引用
收藏
页码:4381 / 4393
页数:13
相关论文
共 50 条
  • [41] Depth Estimation for Small Obstacles Based on Monocular Vision
    Gao, Wang
    Wang, Changqing
    Li, Lei
    PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), 2020, : 407 - 411
  • [42] Swin-Depth: Using Transformers and Multi-Scale Fusion for Monocular-Based Depth Estimation
    Cheng, Zeyu
    Zhang, Yi
    Tang, Chengkai
    IEEE SENSORS JOURNAL, 2021, 21 (23) : 26912 - 26920
  • [43] Monocular Depth Estimation Method Based on Plane Coefficient Representation with Adaptive Depth Distribution
    Wang Jiajun
    Liu Yue
    Wu Yuhui
    Sha Hao
    Wang Yongtian
    ACTA OPTICA SINICA, 2023, 43 (14)
  • [44] The Constraints between Edge Depth and Uncertainty for Monocular Depth Estimation
    Wu, Shouying
    Li, Wei
    Liang, Binbin
    Huang, Guoxin
    ELECTRONICS, 2021, 10 (24)
  • [45] A Deep Joint Network for Monocular Depth Estimation Based on Pseudo-Depth Supervision
    Tan, Jiahai
    Gao, Ming
    Duan, Tao
    Gao, Xiaomei
    MATHEMATICS, 2023, 11 (22)
  • [46] DTTNet: Depth Transverse Transformer Network for Monocular Depth Estimation
    Kamath, Shreyas K. M.
    Rajeev, Srijith
    Panetta, Karen
    Agaian, Sos S.
    MULTIMODAL IMAGE EXPLOITATION AND LEARNING 2022, 2022, 12100
  • [47] ENTROPY CRITERION FOR PROGRESSIVE LAPLACIAN PYRAMID-BASED IMAGE TRANSMISSION
    MONGATTI, G
    ALPARONE, L
    BARONTI, S
    ELECTRONICS LETTERS, 1989, 25 (07) : 450 - 451
  • [48] Depth-Relative Self Attention for Monocular Depth Estimation
    Shim, Kyuhong
    Kim, Jiyoung
    Lee, Gusang
    Shim, Byonghyo
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 1396 - 1404
  • [49] Laplacian pyramid-based change detection in multitemporal SAR images
    Geetha, R. Vijaya
    Kalaivani, S.
    EUROPEAN JOURNAL OF REMOTE SENSING, 2019, 52 (01) : 463 - 483
  • [50] Dense monocular depth estimation for stereoscopic vision based on pyramid transformer and multi-scale feature fusion
    Xia, Zhongyi
    Wu, Tianzhao
    Wang, Zhuoyan
    Zhou, Man
    Wu, Boqi
    Chan, C. Y.
    Kong, Ling Bing
    SCIENTIFIC REPORTS, 2024, 14 (01)