MARS-LVIG dataset: A multi-sensor aerial robots SLAM dataset for LiDAR-visual-inertial-GNSS fusion

被引:7
|
作者
Li, Haotian [1 ]
Zou, Yuying [1 ]
Chen, Nan [1 ]
Lin, Jiarong [1 ]
Liu, Xiyuan [1 ]
Xu, Wei [1 ]
Zheng, Chunran [1 ]
Li, Rundong [1 ]
He, Dongjiao [1 ]
Kong, Fanze [1 ]
Cai, Yixi [1 ]
Liu, Zheng [1 ]
Zhou, Shunbo [2 ]
Xue, Kaiwen [2 ]
Zhang, Fu [1 ,3 ]
机构
[1] Univ Hong Kong, Dept Mech Engn, Hong Kong, Peoples R China
[2] Huawei Cloud Comp Technol Co Ltd, Huawei Cloud Comp Tech Innovat Dept, Guian, Peoples R China
[3] Univ Hong Kong, Dept Mechan Engn, Mechatron & Robot Syst Lab, Pokfulam, HW 7-18, Hong Kong, Peoples R China
来源
关键词
Dataset; aerial robots; multi-sensor fusion; LiDAR; camera; Simultaneous Localization and Mapping; Global Navigation Satellite System; Inertial Measurement Unit; URBAN DATASET; ROBUST; LOCALIZATION; BENCHMARK; ODOMETRY; CAMERA; READY;
D O I
10.1177/02783649241227968
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
In recent years, advancements in Light Detection and Ranging (LiDAR) technology have made 3D LiDAR sensors more compact, lightweight, and affordable. This progress has spurred interest in integrating LiDAR with sensors such as Inertial Measurement Units (IMUs) and cameras for Simultaneous Localization and Mapping (SLAM) research. Public datasets covering different scenarios, platforms, and viewpoints are crucial for multi-sensor fusion SLAM studies, yet most focus on handheld or vehicle-mounted devices with front or 360-degree views. Data from aerial vehicles with downward-looking views is scarce, existing relevant datasets usually feature low altitudes and are mostly limited to small campus environments. To fill this gap, we introduce the Multi-sensor Aerial Robots SLAM dataset (MARS-LVIG dataset), providing unique aerial downward-looking LiDAR-Visual-Inertial-GNSS data with viewpoints from altitudes between 80 m and 130 m. The dataset not only offers new aspects to test and evaluate existing SLAM algorithms, but also brings new challenges which can facilitate researches and developments of more advanced SLAM algorithms. The MARS-LVIG dataset contains 21 sequences, acquired across diversified large-area environments including an aero-model airfield, an island, a rural town, and a valley. Within these sequences, the UAV has speeds varying from 3 m/s to 12 m/s, a scanning area reaching up to 577,000 m(2), and the max path length of 7.148 km in a single flight. This dataset encapsulates data collected by a lightweight, hardware-synchronized sensor package that includes a solid-state 3D LiDAR, a global-shutter RGB camera, IMUs, and a raw message receiver of the Global Navigation Satellite System (GNSS). For algorithm evaluation, this dataset releases ground truth of both localization and mapping, which are acquired by on-board Real-time Kinematic (RTK) and DJI L1 (post-processed by its supporting software DJI Terra), respectively. The dataset can be downloaded from: https://mars.hku.hk/dataset.html.
引用
收藏
页码:1114 / 1127
页数:14
相关论文
共 34 条
  • [1] A multi-sensor fusion SLAM algorithm for indoor aerial robots
    Lin, Xumei
    Zai, Weiqiang
    Lin, Qihang
    Zhang, Qinghua
    JOURNAL OF CONTROL AND DECISION, 2024,
  • [2] Multi-Sensor Fusion for Aerial Robots in Industrial GNSS-Denied Environments
    Carrasco, Paloma
    Cuesta, Francisco
    Caballero, Rafael
    Perez-Grau, Francisco J.
    Viguria, Antidio
    APPLIED SCIENCES-BASEL, 2021, 11 (09):
  • [3] M2DGR: A Multi-Sensor and Multi-Scenario SLAM Dataset for Ground Robots
    Yin, Jie
    Li, Ang
    Li, Tao
    Yu, Wenxian
    Zou, Danping
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02): : 2266 - 2273
  • [4] A multi-sensor fusion SLAM approach for mobile robots
    Fang, Fang
    Ma, Xudong
    Dai, Xianzhong
    2005 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS AND AUTOMATIONS, VOLS 1-4, CONFERENCE PROCEEDINGS, 2005, : 1837 - 1841
  • [5] A Robust Lidar SLAM System Based on Multi-Sensor Fusion
    Zhang, Fubin
    Zhang, Bingshuo
    Sun, Chenghao
    2022 11TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND INFORMATION SCIENCES (ICCAIS), 2022, : 130 - 135
  • [6] An Embedded High-Precision GNSS-Visual-Inertial Multi-Sensor Fusion Suite
    Liu, Cheng
    Xiong, Shuai
    Geng, Yongchao
    Cheng, Song
    Hu, Fang
    Shao, Bo
    Li, Fang
    Zhang, Jie
    NAVIGATION-JOURNAL OF THE INSTITUTE OF NAVIGATION, 2023, 70 (04): : 1 - 23
  • [7] Camera, LiDAR, and IMU Based Multi-Sensor Fusion SLAM: A Survey
    Zhu, Jun
    Li, Hongyi
    Zhang, Tao
    TSINGHUA SCIENCE AND TECHNOLOGY, 2024, 29 (02): : 415 - 429
  • [8] USTC FLICAR: A sensors fusion dataset of LiDAR-inertial-camera for heavy-duty autonomous aerial work robots
    Wang, Ziming
    Liu, Yujiang
    Duan, Yifan
    Li, Xingchen
    Zhang, Xinran
    Ji, Jianmin
    Dong, Erbao
    Zhang, Yanyong
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2023, 42 (11): : 1015 - 1047
  • [9] Map Construction Based on LiDAR Vision Inertial Multi-Sensor Fusion
    Zhang, Chuanwei
    Lei, Lei
    Ma, Xiaowen
    Zhou, Rui
    Shi, Zhenghe
    Guo, Zhongyu
    WORLD ELECTRIC VEHICLE JOURNAL, 2021, 12 (04):
  • [10] NTU VIRAL: A visual-inertial-ranging-lidar dataset, from an aerial vehicle viewpoint
    Nguyen, Thien-Minh
    Yuan, Shenghai
    Cao, Muqing
    Lyu, Yang
    Nguyen, Thien H.
    Xie, Lihua
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2022, 41 (03): : 270 - 280