Scale-Aware Visual-Inertial Depth Estimation and Odometry Using Monocular Self-Supervised Learning

被引:3
|
作者
Lee, Chungkeun [1 ]
Kim, Changhyeon [2 ]
Kim, Pyojin [3 ]
Lee, Hyeonbeom [4 ]
Kim, H. Jin [5 ]
机构
[1] Seoul Natl Univ, Inst Adv Aerosp Technol, Seoul 08826, South Korea
[2] Seoul Natl Univ, Automation & Syst Res Inst, Seoul 08826, South Korea
[3] Sookmyung Womens Univ, Dept Mech Syst Engn, Seoul 04312, South Korea
[4] Kyungpook Natl Univ, Sch Elect & Elect Engn, Daegu 37224, South Korea
[5] Seoul Natl Univ, Dept Mech & Aerosp Engn, Seoul 08826, South Korea
基金
新加坡国家研究基金会;
关键词
Odometry; Deep learning; Loss measurement; Depth measurement; Cameras; Self-supervised learning; Coordinate measuring machines; monocular depth estimation; self-supervised learning; visual-inertial odometry;
D O I
10.1109/ACCESS.2023.3252884
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
For real-world applications with a single monocular camera, scale ambiguity is an important issue. Because self-supervised data-driven approaches that do not require additional data containing scale information cannot avoid the scale ambiguity, state-of-the-art deep-learning-based methods address this issue by learning the scale information from additional sensor measurements. In that regard, inertial measurement unit (IMU) is a popular sensor for various mobile platforms due to its lightweight and inexpensiveness. However, unlike supervised learning that can learn the scale from the ground-truth information, learning the scale from IMU is challenging in a self-supervised setting. We propose a scale-aware monocular visual-inertial depth estimation and odometry method with end-to-end training. To learn the scale from the IMU measurements with end-to-end training in the monocular self-supervised setup, we propose a new loss function named as preintegration loss function, which trains scale-aware ego-motion by comparing the ego-motion integrated from IMU measurement and predicted ego-motion. Since the gravity and the bias should be compensated to obtain the ego-motion by integrating IMU measurements, we design a network to predict the gravity and the bias in addition to the ego-motion and the depth map. The overall performance of the proposed method is compared to state-of-the-art methods in the popular outdoor driving dataset, i.e., KITTI dataset, and the author-collected indoor driving dataset. In the KITTI dataset, the proposed method shows competitive performance compared with state-of-the-art monocular depth estimation and odometry methods, i.e., root-mean-square error of 5.435 m in the KITTI Eigen split and absolute trajectory error of 22.46 m and 0.2975 degrees in the KITTI odometry 09 sequence. Different from other up-to-scale monocular methods, the proposed method can estimate the metric-scaled depth and camera poses. Additional experiments on the author-collected indoor driving dataset qualitatively confirm the accurate performance of metric-depth and metric pose estimations.
引用
收藏
页码:24087 / 24102
页数:16
相关论文
共 50 条
  • [1] SelfVIO: Self-supervised deep monocular Visual-Inertial Odometry and depth estimation
    Almalioglu, Yasin
    Turan, Mehmet
    Saputra, Muhamad Risqi U.
    de Gusmao, Pedro P. B.
    Markham, Andrew
    Trigoni, Niki
    NEURAL NETWORKS, 2022, 150 : 119 - 136
  • [2] A self-supervised monocular odometry with visual-inertial and depth representations
    Zhao, Lingzhe
    Xiang, Tianyu
    Wang, Zhuping
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2024, 361 (06):
  • [3] Self-Supervised Learning of Monocular Visual Odometry and Depth with Uncertainty-Aware Scale Consistency
    Zhang, Guanwen (guanwen.zh@nwpu.edu.cn), 1600, Institute of Electrical and Electronics Engineers Inc.
  • [4] Self-supervised Scale Recovery for Decoupled Visual-inertial Odometry
    Zhang, Tianyu
    Zhu, Dongchen
    Shi, Wenjun
    Liu, Yanqing
    Zhang, Guanghui
    Zhang, Xiaolin
    Li, Jiamao
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (02) : 1612 - 1619
  • [5] Scale-Aware Monocular Visual-Inertial Pose Estimation for Aerial Robots
    Xiong, Minjun
    Lu, Huimin
    Xiong, Dan
    Xiao, Junhao
    Lv, Ming
    2017 CHINESE AUTOMATION CONGRESS (CAC), 2017, : 7030 - 7034
  • [6] Enhancing self-supervised monocular depth estimation with traditional visual odometry
    Andraghetti, Lorenzo
    Myriokefalitakis, Panteleimon
    Dovesi, Pier Luigi
    Luque, Belen
    Poggi, Matteo
    Pieropan, Alessandro
    Mattoccia, Stefano
    2019 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2019), 2019, : 424 - 433
  • [7] A Self-Supervised, Differentiable Kalman Filter for Uncertainty-Aware Visual-Inertial Odometry
    Wagstaff, Brandon
    Wise, Emmett
    Kelly, Jonathan
    2022 IEEE/ASME INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS (AIM), 2022, : 1388 - 1395
  • [8] Self-supervised multi-frame depth estimation with visual-inertial pose transformer and monocular guidance
    Wang, Xiang
    Luo, Haonan
    Wang, Zihang
    Zheng, Jin
    Bai, Xiao
    INFORMATION FUSION, 2024, 108
  • [9] Self-supervised Monocular Depth and Visual Odometry Learning with Scale-consistent Geometric Constraints
    Xiong, Mingkang
    Zhang, Zhenghong
    Zhong, Weilin
    Ji, Jinsheng
    Liu, Jiyuan
    Xiong, Huilin
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 963 - 969
  • [10] Self-supervised deep monocular visual odometry and depth estimation with observation variation
    Zhao, Wentao
    Wang, Yanbo
    Wang, Zehao
    Li, Rui
    Xiao, Peng
    Wang, Jingchuan
    Guo, Rui
    DISPLAYS, 2023, 80