Graph attention network-optimized dynamic monocular visual odometry

被引:2
|
作者
Hongru, Zhao [1 ]
Xiuquan, Qiao [1 ]
机构
[1] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
关键词
Monocular visual odometry; Multi-task learning; Multi-view geometry; Dynamic objects removal; Graph attention network; SEMANTIC SEGMENTATION;
D O I
10.1007/s10489-023-04687-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Monocular Visual Odometry (VO) is often formulated as a sequential dynamics problem that relies on scene rigidity assumption. One of the main challenges is rejecting moving objects and estimating camera pose in dynamic environments. Existing methods either take the visual cues in the whole image equally or eliminate the fixed semantic categories by heuristics or attention mechanisms. However, they fail to tackle unknown dynamic objects which are not labeled in the training sets of the network. To solve these issues, we propose a novel framework, named graph attention network (GAT)-optimized dynamic monocular visual odometry (GDM-VO), to remove dynamic objects explicitly with semantic segmentation and multi-view geometry in this paper. Firstly, we employ a multi-task learning network to perform semantic segmentation and depth estimation. Then, we reject priori known and unknown objective moving objects through semantic information and multi-view geometry, respectively. Furthermore, to our best knowledge, we are the first to leverage GAT to capture long-range temporal dependencies from consecutive image sequences adaptively, while existing sequential modeling approaches need to select information manually. Extensive experiments on the KITTI and TUM datasets demonstrate the superior performance of GDM-VO overs existing state-of-the-art classical and learning-based monocular VO.
引用
收藏
页码:23067 / 23082
页数:16
相关论文
共 50 条
  • [31] Resolving Scale Ambiguity for Monocular Visual Odometry
    Choi, Sunglok
    Park, Jaehyun
    Yu, Wonpil
    2013 10TH INTERNATIONAL CONFERENCE ON UBIQUITOUS ROBOTS AND AMBIENT INTELLIGENCE (URAI), 2013, : 604 - 608
  • [32] Deep Monocular Visual Odometry for Ground Vehicle
    Wang, Xiangwei
    Zhang, Hui
    IEEE ACCESS, 2020, 8 : 175220 - 175229
  • [33] Multimodal Scale Estimation for Monocular Visual Odometry
    Fanani, Nolang
    Stuerck, Alina
    Barnada, Marc
    Mester, Rudolf
    2017 28TH IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV 2017), 2017, : 1714 - 1721
  • [34] Monocular Visual Odometry for underground railway scenarios
    Etxeberria-Garcia, Mikel
    Labayen, Mikel
    Eizaguirre, Fernando
    Zamalloa, Maider
    Arana-Arexolaleiba, Nestor
    FIFTEENTH INTERNATIONAL CONFERENCE ON QUALITY CONTROL BY ARTIFICIAL VISION, 2021, 11794
  • [35] LIMO: Lidar-Monocular Visual Odometry
    Graeter, Johannes
    Wilczynski, Alexander
    Lauer, Martin
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 7872 - 7879
  • [36] Perceptual Enhancement for Unsupervised Monocular Visual Odometry
    Wang, Zhongyi
    Shen, Mengjiao
    Liu, Chengju
    Chen, Qijun
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2025, 23 (01) : 346 - 357
  • [37] Milk: Monocular Visual Odometry with Motion Constraints
    Choi, Sunglok
    Yu, Wonpil
    2012 9TH INTERNATIONAL CONFERENCE ON UBIQUITOUS ROBOTS AND AMBIENT INTELLIGENCE (URAL), 2012, : 199 - 199
  • [38] Robust Monocular Visual Odometry by Uncertainty Voting
    Van Hamme, David
    Veelaert, Peter
    Philips, Wilfried
    2011 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2011, : 643 - 647
  • [39] STMVO: biologically inspired monocular visual odometry
    Yangming Li
    Jian Zhang
    Shuai Li
    Neural Computing and Applications, 2018, 29 : 215 - 225
  • [40] STMVO: biologically inspired monocular visual odometry
    Li, Yangming
    Zhang, Jian
    Li, Shuai
    NEURAL COMPUTING & APPLICATIONS, 2018, 29 (06): : 215 - 225