Improving Monocular Depth Estimation by Semantic Pre-training

被引:1
|
作者
Rottmann, Peter [1 ]
Posewsky, Thorbjorn [1 ,2 ]
Milioto, Andres [1 ]
Stachniss, Cyrill [1 ]
Behley, Jens [1 ]
机构
[1] Univ Bonn, Bonn, Germany
[2] Ibeo Automot Syst GmbH, Hamburg, Germany
关键词
D O I
10.1109/IROS51168.2021.9636546
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Knowing the distance to nearby objects is crucial for autonomous cars to navigate safely in everyday traffic. In this paper, we investigate monocular depth estimation, which advanced substantially within the last years and is providing increasingly more accurate results while only requiring a single camera image as input. In line with recent work, we use an encoder-decoder structure with so-called packing layers to estimate depth values in a self-supervised fashion. We propose integrating a joint pre-training of semantic segmentation plus depth estimation on a dataset providing semantic labels. By using a separate semantic decoder that is only needed for pre-training, we can keep the network comparatively small. Our extensive experimental evaluation shows that the addition of such pre-training improves the depth estimation performance substantially. Finally, we show that we achieve competitive performance on the KITTI dataset despite using a much smaller and more efficient network.
引用
收藏
页码:5916 / 5923
页数:8
相关论文
共 50 条
  • [21] Graph semantic information for self-supervised monocular depth estimation
    Zhang, Dongdong
    Wang, Chunping
    Wang, Huiying
    Fu, Qiang
    PATTERN RECOGNITION, 2024, 156
  • [22] Semi-Supervised Monocular Depth Estimation Based on Semantic Supervision
    Min Yue
    Guangyuan Fu
    Ming Wu
    Hongqiao Wang
    Journal of Intelligent & Robotic Systems, 2020, 100 : 455 - 463
  • [23] Improving AMR Parsing with Sequence-to-Sequence Pre-training
    Xu, Dongqin
    Li, Junhui
    Zhu, Muhua
    Min Zhang
    Zhou, Guodong
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 2501 - 2511
  • [24] Improving Knowledge Tracing via Pre-training Question Embeddings
    Liu, Yunfei
    Yang, Yang
    Chen, Xianyu
    Shen, Jian
    Zhang, Haifeng
    Yu, Yong
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 1577 - 1583
  • [25] VarCLR: Variable Semantic Representation Pre-training via Contrastive Learning
    Chen, Qibin
    Lacomis, Jeremy
    Schwartz, Edward J.
    Neubig, Graham
    Vasilescu, Bogdan
    Le Goues, Claire
    2022 ACM/IEEE 44TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE 2022), 2022, : 2327 - 2339
  • [26] Evaluating the Use of Synthetic Queries for Pre-training a Semantic Query Tagger
    Bassani, Elias
    Pasi, Gabriella
    ADVANCES IN INFORMATION RETRIEVAL, PT II, 2022, 13186 : 39 - 46
  • [27] Improving negation detection with negation-focused pre-training
    Hung Thinh Truong
    Baldwin, Timothy
    Cohn, Trevor
    Verspoor, Karin
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 4188 - 4193
  • [28] Unsupervised Pre-training on Improving the Performance of Neural Network in Regression
    Salida, Pallabi
    Vij, Prateek
    Baruah, Rashmi Dutta
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [29] Improving Question Answering by Commonsense-Based Pre-training
    Zhong, Wanjun
    Tang, Duyu
    Duan, Nan
    Zhou, Ming
    Wang, Jiahai
    Yin, Jian
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING (NLPCC 2019), PT I, 2019, 11838 : 16 - 28
  • [30] Self-supervised Pre-training for Semantic Segmentation in an Indoor Scene
    Shrestha, Sulabh
    Li, Yimeng
    Kosecka, Jana
    2024 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WORKSHOPS, WACVW 2024, 2024, : 625 - 635