Pyramid frequency network with spatial attention residual refinement module for monocular depth estimation

被引:9
|
作者
Lu, Zhengyang [1 ]
Chen, Ying [1 ]
机构
[1] Jiangnan Univ, Key Lab Adv Proc Control Light Ind, Minist Educ, Wuxi, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
monocular depth estimation; three-dimensional reconstruction; frequency domain; convolutional neural network;
D O I
10.1117/1.JEI.31.2.023005
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep-learning-based approaches to depth estimation are rapidly advancing, offering superior performance over existing methods. To estimate the depth in real-world scenarios, depth estimation models require the robustness of various noise environments. We propose a pyramid frequency network (PFN) with spatial attention residual refinement module (SARRM) to deal with the weak robustness of existing deep-learning methods. To reconstruct depth maps with accurate details, the SARRM constructs a residual fusion method with an attention mechanism to refine the blur depth. The frequency division strategy is designed, and the frequency pyramid network is developed to extract features from multiple frequency bands. With the frequency strategy, PFN achieves better visual accuracy than state-of-the-art methods in both indoor and outdoor scenes on Make3D, KITTI depth, and NYUv2 datasets. Additional experiments on the noisy NYUv2 dataset demonstrate that PFN is more reliable than existing deep-learning methods in high noise scenes. (C) 2022 SPIE and IS&T
引用
收藏
页数:18
相关论文
共 50 条
  • [31] Dermoscopic image segmentation based on Pyramid Residual Attention Module
    Jiang, Yun
    Cheng, Tongtong
    Dong, Jinkun
    Liang, Jing
    Zhang, Yuan
    Lin, Xin
    Yao, Huixia
    [J]. PLOS ONE, 2022, 17 (09):
  • [32] LAM-Depth: Laplace-Attention Module-Based Self-Supervised Monocular Depth Estimation
    Wei, Jiansheng
    Pan, Shuguo
    Gao, Wang
    Guo, Peng
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024,
  • [33] Efficient unsupervised monocular depth estimation using attention guided generative adversarial network
    Sumanta Bhattacharyya
    Ju Shen
    Stephen Welch
    Chen Chen
    [J]. Journal of Real-Time Image Processing, 2021, 18 : 1357 - 1368
  • [34] Efficient unsupervised monocular depth estimation using attention guided generative adversarial network
    Bhattacharyya, Sumanta
    Shen, Ju
    Welch, Stephen
    Chen, Chen
    [J]. JOURNAL OF REAL-TIME IMAGE PROCESSING, 2021, 18 (04) : 1357 - 1368
  • [35] Self-supervised coarse-to-fine monocular depth estimation using a lightweight attention module
    Yuanzhen Li
    Fei Luo
    Chunxia Xiao
    [J]. Computational Visual Media, 2022, 8 : 631 - 647
  • [36] Triaxial Squeeze Attention Module and Mutual-Exclusion Loss Based Unsupervised Monocular Depth Estimation
    Wei, Jiansheng
    Pan, Shuguo
    Gao, Wang
    Zhao, Tao
    [J]. NEURAL PROCESSING LETTERS, 2022, 54 (05) : 4375 - 4390
  • [37] Trap Attention: Monocular Depth Estimation with Manual Traps
    Ning, Chao
    Gan, Hongping
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 5033 - 5043
  • [38] Self-supervised coarse-to-fine monocular depth estimation using a lightweight attention module
    Li, Yuanzhen
    Luo, Fei
    Xiao, Chunxia
    [J]. COMPUTATIONAL VISUAL MEDIA, 2022, 8 (04) : 631 - 647
  • [39] Triaxial Squeeze Attention Module and Mutual-Exclusion Loss Based Unsupervised Monocular Depth Estimation
    Jiansheng Wei
    Shuguo Pan
    Wang Gao
    Tao Zhao
    [J]. Neural Processing Letters, 2022, 54 : 4375 - 4390
  • [40] CATNet: Convolutional attention and transformer for monocular depth estimation
    Tang, Shuai
    Lu, Tongwei
    Liu, Xuanxuan
    Zhou, Huabing
    Zhang, Yanduo
    [J]. PATTERN RECOGNITION, 2024, 145