Multi-Scale Spatial Attention-Guided Monocular Depth Estimation With Semantic Enhancement

被引:15
|
作者
Xu, Xianfa [1 ]
Chen, Zhe [1 ]
Yin, Fuliang [1 ]
机构
[1] Dalian Univ Technol, Sch Informat & Commun Engn, Dalian 116023, Peoples R China
关键词
Estimation; Semantics; Mutual information; Feature extraction; Correlation; Cameras; Visualization; Depth estimation; multi-scale spatial attention-guided; mutual information; semantic enhancement;
D O I
10.1109/TIP.2021.3120670
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Depth estimation from single monocular image is a vital but challenging task in 3D vision and scene understanding. Previous unsupervised methods have yielded impressive results, but the predicted depth maps still have several disadvantages such as missing small objects and object edge blurring. To address these problems, a multi-scale spatial attention guided monocular depth estimation method with semantic enhancement is proposed. Specifically, we first construct a multi-scale spatial attention-guided block based on atrous spatial pyramid pooling and spatial attention. Then, the correlation between the left and right views is fully explored by mutual information to obtain a more robust feature representation. Finally, we design a double-path prediction network to simultaneously generate depth maps and semantic labels. The proposed multi-scale spatial attention-guided block can focus more on the objects, especially on small objects. Moreover, the additional semantic information also enables the objects edge in the predicted depth maps more sharper. We conduct comprehensive evaluations on public benchmark datasets, such as KITTI and Make3D. The experiment results well demonstrate the effectiveness of the proposed method and achieve better performance than other self-supervised methods.
引用
收藏
页码:8811 / 8822
页数:12
相关论文
共 50 条
  • [1] Attention-Guided Multi-Scale Fusion Network for Similar Objects Semantic Segmentation
    Yao, Fengqin
    Wang, Shengke
    Ding, Laihui
    Zhong, Guoqiang
    Li, Shu
    Xu, Zhiwei
    [J]. COGNITIVE COMPUTATION, 2024, 16 (01) : 366 - 376
  • [2] Attention-Guided Multi-Scale Fusion Network for Similar Objects Semantic Segmentation
    Fengqin Yao
    Shengke Wang
    Laihui Ding
    Guoqiang Zhong
    Shu Li
    Zhiwei Xu
    [J]. Cognitive Computation, 2024, 16 : 366 - 376
  • [3] Multi-scale Residual Pyramid Attention Network for Monocular Depth Estimation
    Liu, Jing
    Zhang, Xiaona
    Li, Zhaoxin
    Mao, Tianlu
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 5137 - 5144
  • [4] GLIMS: Attention-guided lightweight multi-scale hybrid network for volumetric semantic segmentation
    Yazici, Ziya Ata
    Oksuz, Ilkay
    Ekenel, Hazim Kemal
    [J]. IMAGE AND VISION COMPUTING, 2024, 146
  • [5] Multi-Scale Attention-Guided Network for mammograms classification
    Xu, Chunbo
    Lou, Meng
    Qi, Yunliang
    Wang, Yiming
    Pi, Jiande
    Ma, Yide
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2021, 68
  • [6] Lightweight multi-scale attention-guided network for real-time semantic segmentation
    Hu, Xuegang
    Liu, Yuanjing
    [J]. IMAGE AND VISION COMPUTING, 2023, 139
  • [7] Monocular Image Depth Estimation Based on Multi-Scale Attention Oriented Network
    Liu, Jieping
    Wen, Junwen
    Liang, Yaling
    [J]. Huanan Ligong Daxue Xuebao/Journal of South China University of Technology (Natural Science), 2020, 48 (12): : 52 - 62
  • [8] Multi-scale depth classification network for monocular depth estimation
    Yang, Yi
    Tian, Lihua
    Li, Chen
    Zhang, Botong
    [J]. COMPUTERS & ELECTRICAL ENGINEERING, 2022, 102
  • [9] Monocular Depth Estimation With Multi-Scale Feature Fusion
    Xu, Xianfa
    Chen, Zhe
    Yin, Fuliang
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 678 - 682
  • [10] Monocular depth estimation with multi-scale feature fusion
    Wang, Quande
    Zhang, Songtao
    [J]. Huazhong Keji Daxue Xuebao (Ziran Kexue Ban)/Journal of Huazhong University of Science and Technology (Natural Science Edition), 2020, 48 (05): : 7 - 12