Monocular Depth Estimation with Adaptive Geometric Attention

被引:8
|
作者
Naderi, Taher [1 ]
Sadovnik, Amir [1 ]
Hayward, Jason [2 ]
Qi, Hairong [1 ]
机构
[1] Univ Tennessee, Dept Elect Engn & Comp Sci, Knoxville, TN 37996 USA
[2] Univ Tennessee, Dept Nucl Engn, Knoxville, TN 37996 USA
关键词
SEGMENTATION; NETWORK;
D O I
10.1109/WACV51458.2022.00069
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Single image depth estimation is an ill-posed problem. That is, it is not mathematically possible to uniquely estimate the 3rd dimension (or depth) from a single 2D image. Hence, additional constraints need to be incorporated in order to regulate the solution space. In this paper, we explore the idea of constraining the model by taking advantage of the similarity between the RGB image and the corresponding depth map at the geometric edges of the 3D scene for more accurate depth estimation. We propose a general light-weight adaptive geometric attention module that uses the cross-correlation between the encoder and the decoder as a measure of this similarity. More precisely, we use the cosine similarity between the local embedded features in the encoder and the decoder at each spatial point. The proposed module along with the encoder-decoder network is trained in an end-to-end fashion and achieves superior and competitive performance in comparison with other state-of-the-art methods. In addition, adding our module to the base encoder-decoder model adds only an additional 0.03% (or 0.0003) parameters. Therefore, this module can be added to any base encoder-decoder network without changing its structure to address any task at hand.
引用
收藏
页码:617 / 627
页数:11
相关论文
共 50 条
  • [1] Geometric Pretraining for Monocular Depth Estimation
    Wang, Kaixuan
    Chen, Yao
    Guo, Hengkai
    Wen, Linfu
    Shen, Shaojie
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 4782 - 4788
  • [2] EdgeConv with Attention Module for Monocular Depth Estimation
    Lee, Minhyeok
    Hwang, Sangwon
    Park, Chaewon
    Lee, Sangyoun
    [J]. 2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 2364 - 2373
  • [3] Bidirectional Attention Network for Monocular Depth Estimation
    Aich, Shubhra
    Vianney, Jean Marie Uwabeza
    Islam, Md Amirul
    Kaur, Mannat
    Liu, Bingbing
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 11746 - 11752
  • [4] Depth-Relative Self Attention for Monocular Depth Estimation
    Shim, Kyuhong
    Kim, Jiyoung
    Lee, Gusang
    Shim, Byonghyo
    [J]. PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 1396 - 1404
  • [5] Adaptive confidence thresholding for monocular depth estimation
    Choi, Hyesong
    Lee, Hunsang
    Kim, Sunkyung
    Kim, Sunok
    Kim, Seungryong
    Sohn, Kwanghoon
    Min, Dongbo
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 12788 - 12798
  • [6] SaccadeCam: Adaptive Visual Attention for Monocular Depth Sensing
    Tilmon, Brevin
    Koppal, Sanjeev J.
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 5989 - 5998
  • [7] Trap Attention: Monocular Depth Estimation with Manual Traps
    Ning, Chao
    Gan, Hongping
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 5033 - 5043
  • [8] Unsupervised Monocular Depth Estimation With Channel and Spatial Attention
    Wang, Zhuping
    Dai, Xinke
    Guo, Zhanyu
    Huang, Chao
    Zhang, Hao
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (06) : 7860 - 7870
  • [9] CATNet: Convolutional attention and transformer for monocular depth estimation
    Tang, Shuai
    Lu, Tongwei
    Liu, Xuanxuan
    Zhou, Huabing
    Zhang, Yanduo
    [J]. PATTERN RECOGNITION, 2024, 145
  • [10] Attention Mechanism Used in Monocular Depth Estimation: An Overview
    Li, Yundong
    Wei, Xiaokun
    Fan, Hanlu
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (17):