Online supervised attention-based recurrent depth estimation from monocular video

被引:0
|
作者
Maslov D. [1 ]
Makarov I. [1 ,2 ]
机构
[1] School of Data Analysis and Artificial Intelligence, HSE University, Moscow
[2] Samsung-PDMI Joint AI Center, St. Petersburg Department of Steklov Institute of Mathematics, St. Petersburg
来源
Maslov, Dmitrii (dvmaslov@edu.hse.ru) | 1600年 / PeerJ Inc.卷 / 06期
关键词
Augmented Reality; Autonomous Vehicles; Computer Science Methods; Computer Vision; Deep Convolutional Neural Networks; Depth Reconstruction; Recurrent Neural Networks;
D O I
10.7717/PEERJ-CS.317
中图分类号
学科分类号
摘要
Autonomous driving highly depends on depth information for safe driving. Recently, major improvements have been taken towards improving both supervised and self-supervised methods for depth reconstruction. However, most of the current approaches focus on single frame depth estimation, where quality limit is hard to beat due to limitations of supervised learning of deep neural networks in general. One of the way to improve quality of existing methods is to utilize temporal information from frame sequences. In this paper, we study intelligent ways of integrating recurrent block in common supervised depth estimation pipeline. We propose a novel method, which takes advantage of the convolutional gated recurrent unit (convGRU) and convolutional long short-term memory (convLSTM). We compare use of convGRU and convLSTM blocks and determine the best model for real-time depth estimation task. We carefully study training strategy and provide new deep neural networks architectures for the task of depth estimation from monocular video using information from past frames based on attention mechanism. We demonstrate the efficiency of exploiting temporal information by comparing our best recurrent method with existing image-based and video-based solutions for monocular depth reconstruction. © 2020. Maslov and Makarov. All Rights Reserved.
引用
收藏
页码:1 / 22
页数:21
相关论文
共 50 条
  • [21] Attention-based framework for weakly supervised video anomaly detection
    Hualin Ma
    Liyan Zhang
    The Journal of Supercomputing, 2022, 78 : 8409 - 8429
  • [22] MDSNet: self-supervised monocular depth estimation for video sequences using self-attention and threshold mask
    Zhao, Jiaqi
    Zhao, Chaoyue
    Liu, Chunling
    Zhang, Chaojian
    Zhang, Wang
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (05)
  • [23] Attention-based framework for weakly supervised video anomaly detection
    Ma, Hualin
    Zhang, Liyan
    JOURNAL OF SUPERCOMPUTING, 2022, 78 (06): : 8409 - 8429
  • [24] LAM-Depth: Laplace-Attention Module-Based Self-Supervised Monocular Depth Estimation
    Wei, Jiansheng
    Pan, Shuguo
    Gao, Wang
    Guo, Peng
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (10) : 13706 - 13716
  • [25] Depth estimation of supervised monocular images based on semantic segmentation
    Wang, Qi
    Piao, Yan
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 90
  • [26] Dual-attention-based semantic-aware self-supervised monocular depth estimation
    Xu, Jinze
    Ye, Feng
    Lai, Yizong
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (24) : 65579 - 65601
  • [27] Global Pose Estimation with an Attention-based Recurrent Network
    Parisotto, Emilio
    Chaplot, Devendra Singh
    Zhang, Jian
    Salakhutdinov, Ruslan
    PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, : 350 - 359
  • [28] ADAADepth: Adapting Data Augmentation and Attention for Self-Supervised Monocular Depth Estimation
    Kaushik, Vinay
    Jindgar, Kartik
    Lall, Brejesh
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (04) : 7791 - 7798
  • [29] Joint Soft-Hard Attention for Self-Supervised Monocular Depth Estimation
    Fan, Chao
    Yin, Zhenyu
    Xu, Fulong
    Chai, Anying
    Zhang, Feiqing
    SENSORS, 2021, 21 (21)
  • [30] Transferring knowledge from monocular completion for self-supervised monocular depth estimation
    Sun, Lin
    Li, Yi
    Liu, Bingzheng
    Xu, Liying
    Zhang, Zhe
    Zhu, Jie
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (29) : 42485 - 42495