Scale-Aware Visual-Inertial Depth Estimation and Odometry Using Monocular Self-Supervised Learning

被引:3
|
作者
Lee, Chungkeun [1 ]
Kim, Changhyeon [2 ]
Kim, Pyojin [3 ]
Lee, Hyeonbeom [4 ]
Kim, H. Jin [5 ]
机构
[1] Seoul Natl Univ, Inst Adv Aerosp Technol, Seoul 08826, South Korea
[2] Seoul Natl Univ, Automation & Syst Res Inst, Seoul 08826, South Korea
[3] Sookmyung Womens Univ, Dept Mech Syst Engn, Seoul 04312, South Korea
[4] Kyungpook Natl Univ, Sch Elect & Elect Engn, Daegu 37224, South Korea
[5] Seoul Natl Univ, Dept Mech & Aerosp Engn, Seoul 08826, South Korea
基金
新加坡国家研究基金会;
关键词
Odometry; Deep learning; Loss measurement; Depth measurement; Cameras; Self-supervised learning; Coordinate measuring machines; monocular depth estimation; self-supervised learning; visual-inertial odometry;
D O I
10.1109/ACCESS.2023.3252884
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
For real-world applications with a single monocular camera, scale ambiguity is an important issue. Because self-supervised data-driven approaches that do not require additional data containing scale information cannot avoid the scale ambiguity, state-of-the-art deep-learning-based methods address this issue by learning the scale information from additional sensor measurements. In that regard, inertial measurement unit (IMU) is a popular sensor for various mobile platforms due to its lightweight and inexpensiveness. However, unlike supervised learning that can learn the scale from the ground-truth information, learning the scale from IMU is challenging in a self-supervised setting. We propose a scale-aware monocular visual-inertial depth estimation and odometry method with end-to-end training. To learn the scale from the IMU measurements with end-to-end training in the monocular self-supervised setup, we propose a new loss function named as preintegration loss function, which trains scale-aware ego-motion by comparing the ego-motion integrated from IMU measurement and predicted ego-motion. Since the gravity and the bias should be compensated to obtain the ego-motion by integrating IMU measurements, we design a network to predict the gravity and the bias in addition to the ego-motion and the depth map. The overall performance of the proposed method is compared to state-of-the-art methods in the popular outdoor driving dataset, i.e., KITTI dataset, and the author-collected indoor driving dataset. In the KITTI dataset, the proposed method shows competitive performance compared with state-of-the-art monocular depth estimation and odometry methods, i.e., root-mean-square error of 5.435 m in the KITTI Eigen split and absolute trajectory error of 22.46 m and 0.2975 degrees in the KITTI odometry 09 sequence. Different from other up-to-scale monocular methods, the proposed method can estimate the metric-scaled depth and camera poses. Additional experiments on the author-collected indoor driving dataset qualitatively confirm the accurate performance of metric-depth and metric pose estimations.
引用
收藏
页码:24087 / 24102
页数:16
相关论文
共 50 条
  • [31] Keypoint Heatmap Guided Self-Supervised Monocular Visual Odometry
    Xiu, Haixin
    Liang, Yiyou
    Zeng, Hui
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2022, 105 (04)
  • [32] Learning Occlusion-aware Coarse-to-Fine Depth Map for Self-supervised Monocular Depth Estimation
    Zhou, Zhengming
    Dong, Qiulei
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 6386 - 6395
  • [33] Multimodal Scale Consistency and Awareness for Monocular Self-Supervised Depth Estimation
    Chawla, Hemang
    Varma, Arnav
    Arani, Elahe
    Zonooz, Bahram
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 5140 - 5146
  • [34] Spatial-Aware Dynamic Lightweight Self-Supervised Monocular Depth Estimation
    Song, Linna
    Shi, Dianxi
    Xia, Jianqiang
    Ouyang, Qianying
    Qiao, Ziteng
    Jin, Songchang
    Yang, Shaowu
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (01): : 883 - 890
  • [35] Absolute Monocular Depth Estimation on Robotic Visual and Kinematics Data via Self-Supervised Learning
    Wei, Ruofeng
    Li, Bin
    Zhong, Fangxun
    Mo, Hangjie
    Dou, Qi
    Liu, Yun-Hui
    Sun, Dong
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, : 1 - 14
  • [36] Depth estimation algorithm of monocular image based on self-supervised learning
    Bai L.
    Liu L.-J.
    Li X.-A.
    Wu S.
    Liu R.-Q.
    Jilin Daxue Xuebao (Gongxueban)/Journal of Jilin University (Engineering and Technology Edition), 2023, 53 (04): : 1139 - 1145
  • [37] Self-Supervised Learning of Monocular Depth Estimation Based on Progressive Strategy
    Wang, Huachun
    Sang, Xinzhu
    Chen, Duo
    Wang, Peng
    Yan, Binbin
    Qi, Shuai
    Ye, Xiaoqian
    Yao, Tong
    IEEE TRANSACTIONS ON COMPUTATIONAL IMAGING, 2021, 7 : 375 - 383
  • [38] SENSE: Self-Evolving Learning for Self-Supervised Monocular Depth Estimation
    Li, Guanbin
    Huang, Ricong
    Li, Haofeng
    You, Zunzhi
    Chen, Weikai
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 439 - 450
  • [39] Self-supervised learning of monocular depth using quantized networks
    Lu, Keyu
    Zeng, Chengyi
    Zeng, Yonghu
    NEUROCOMPUTING, 2022, 488 : 634 - 646
  • [40] Self-supervised learning monocular depth estimation from internet photos
    Lin, Xiaocan
    Li, Nan
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 99