Multimodal Features and Accurate Place Recognition With Robust Optimization for Lidar-Visual-Inertial SLAM

被引:3
|
作者
Zhao, Xiongwei [1 ]
Wen, Congcong [2 ]
Manoj Prakhya, Sai [3 ]
Yin, Hongpei [4 ]
Zhou, Rundong [5 ]
Sun, Yijiao [1 ]
Xu, Jie [6 ]
Bai, Haojie [1 ]
Wang, Yang [1 ]
机构
[1] Harbin Inst Technol Shenzhen, Sch Elect & Informat Engn, Shenzhen 518071, Peoples R China
[2] NYU, Tandon Sch Engn, New York, NY 10012 USA
[3] Huawei Munich Res Ctr, D-80992 Munich, Germany
[4] Guangdong Inst Artificial Intelligence & Adv Comp, Guangzhou 510535, Peoples R China
[5] Harbin Inst Technol, Sch Elect & Informat Engn, Harbin 150001, Peoples R China
[6] Harbin Inst Technol, Sch Mech & Elect Engn, Harbin 150001, Peoples R China
关键词
Laser radar; Simultaneous localization and mapping; Visualization; Feature extraction; Optimization; Robot sensing systems; Three-dimensional displays; 3-D lidar loop closure descriptor; lidar-visual-inertial simultaneous localization and mapping (SLAM) (LVINS); robust iterative optimization; state estimation; two-stage loop detection; LINE SEGMENT DETECTOR; REAL-TIME; DESCRIPTOR;
D O I
10.1109/TIM.2024.3370762
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Lidar-visual-inertial simultaneous localization and mapping (SLAM) (LVINS) provides a compelling solution for accurate and robust state estimation and mapping, integrating complementary information from the multisensor data. However, in the front-end processing of existing LVINS systems, methods based on the visual line feature matching typically suffer from low accuracy and are time consuming. In addition, the back-end optimization of current multisensor fusion SLAM systems is adversely affected by feature association outliers, which constrains further enhancements in localization precision. In the loop closure process, the existing lidar loop closure descriptors, relying primarily on 2-D information from point clouds, often fall short in complex environments. To effectively tackle these challenges, we introduce the multimodal feature-based LVINS framework, abbreviated as MMF-LVINS. Our framework consists of three major innovations. First, we propose a novel coarse-to-fine (CTF) visual line matching method that utilizes geometric descriptor similarity and optical flow verification, substantially improving both efficiency and accuracy of line feature matching. Second, we present a robust iterative optimization approach featuring a newly proposed adaptive loss function. This function is tailored based on the quality of feature association and incorporates graduated nonconvexity, thereby reducing the impact of outliers on system accuracy. Third, to augment the precision of lidar-based loop closure detection, we introduce an innovative 3-D lidar descriptor that captures spatial, height, and intensity information from the point cloud. We also propose a two-stage place recognition module that synergistically combines both visual and this new lidar descriptor, significantly diminishing cumulative drift. Extensive experimental evaluations on six real-world datasets, including EuRoc, KITTI, NCLT, M2DGR, UrbanNav, and UrbanLoco, demonstrate that our MMF-LVINS system achieves superior state estimation accuracy compared with the existing state-of-the-art methods. These experiments also validate the effectiveness of our advanced techniques in visual line matching, robust iterative optimization, and enhanced lidar loop closure detection.
引用
收藏
页码:1 / 1
页数:16
相关论文
共 50 条
  • [21] Accurate Initialization Method for Monocular Visual-Inertial SLAM
    Amrani, Ahderraouf
    Wang, Hesheng
    2019 3RD INTERNATIONAL SYMPOSIUM ON AUTONOMOUS SYSTEMS (ISAS 2019), 2019, : 159 - 164
  • [22] Unified Multi-Modal Landmark Tracking for Tightly Coupled Lidar-Visual-Inertial Odometry
    Wisth, David
    Camurri, Marco
    Das, Sandipan
    Fallon, Maurice
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 1004 - 1011
  • [23] Sequence searching with CNN features for robust and fast visual place recognition
    Bai, Dongdong
    Wang, Chaoqun
    Zhang, Bo
    Yi, Xiaodong
    Yang, Xuejun
    COMPUTERS & GRAPHICS-UK, 2018, 70 : 270 - 280
  • [24] Cross-Modal LiDAR-Visual-Inertial Localization in Prebuilt LiDAR Point Cloud Map Through Direct Projection
    Leng, Jianghao
    Sun, Chao
    Wang, Bo
    Lan, Yungang
    Huang, Zhishuai
    Zhou, Qinyan
    Liu, Jiahao
    Li, Jiajun
    IEEE SENSORS JOURNAL, 2024, 24 (20) : 33022 - 33035
  • [25] Accurate and Robust Teach and Repeat Navigation by Visual Place Recognition: A CNN Approach
    Camara, Luis G.
    Pivonka, Tomas
    Jilek, Martin
    Gabert, Carl
    Kosnar, Karel
    Preucil, Libor
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 6018 - 6024
  • [26] Robust Visual SLAM with Point and Line Features
    Zuo, Xingxing
    Xie, Xiaojia
    Liu, Yong
    Huang, Guoquan
    2017 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2017, : 1775 - 1782
  • [27] Robust Place Recognition using an Imaging Lidar
    Shan, Tixiao
    Englot, Brendan
    Duarte, Fabio
    Ratti, Carlo
    Rus, Daniela
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 5469 - 5475
  • [28] Robust Indoor Visual-Inertial SLAM with Pedestrian Detection
    Zhang, Heng
    Huang, Ran
    Yuan, Liang
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (IEEE-ROBIO 2021), 2021, : 802 - 807
  • [29] LVI-ExC: A Target-free LiDAR-Visual-Inertial Extrinsic Calibration Framework
    Wang, Zhong
    Zhang, Lin
    Shen, Ying
    Zhou, Yicong
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 3319 - 3327
  • [30] Robust LiDAR visual inertial odometry for dynamic scenes
    Peng, Gang
    Cao, Chong
    Chen, Bocheng
    Hu, Lu
    He, Dingxin
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2024, 35 (09)