File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: MARS-LVIG dataset: A multi-sensor aerial robots SLAM dataset for LiDAR-visual-inertial-GNSS fusion

TitleMARS-LVIG dataset: A multi-sensor aerial robots SLAM dataset for LiDAR-visual-inertial-GNSS fusion
Authors
Keywordsaerial robots
camera
Dataset
Global Navigation Satellite System
Inertial Measurement Unit
LiDAR
multi-sensor fusion
Simultaneous Localization and Mapping
Issue Date2024
Citation
International Journal of Robotics Research, 2024, v. 43, n. 8, p. 1114-1127 How to Cite?
AbstractIn recent years, advancements in Light Detection and Ranging (LiDAR) technology have made 3D LiDAR sensors more compact, lightweight, and affordable. This progress has spurred interest in integrating LiDAR with sensors such as Inertial Measurement Units (IMUs) and cameras for Simultaneous Localization and Mapping (SLAM) research. Public datasets covering different scenarios, platforms, and viewpoints are crucial for multi-sensor fusion SLAM studies, yet most focus on handheld or vehicle-mounted devices with front or 360-degree views. Data from aerial vehicles with downward-looking views is scarce, existing relevant datasets usually feature low altitudes and are mostly limited to small campus environments. To fill this gap, we introduce the Multi-sensor Aerial Robots SLAM dataset (MARS-LVIG dataset), providing unique aerial downward-looking LiDAR-Visual-Inertial-GNSS data with viewpoints from altitudes between 80 m and 130 m. The dataset not only offers new aspects to test and evaluate existing SLAM algorithms, but also brings new challenges which can facilitate researches and developments of more advanced SLAM algorithms. The MARS-LVIG dataset contains 21 sequences, acquired across diversified large-area environments including an aero-model airfield, an island, a rural town, and a valley. Within these sequences, the UAV has speeds varying from 3 m/s to 12 m/s, a scanning area reaching up to 577,000 m2, and the max path length of 7.148 km in a single flight. This dataset encapsulates data collected by a lightweight, hardware-synchronized sensor package that includes a solid-state 3D LiDAR, a global-shutter RGB camera, IMUs, and a raw message receiver of the Global Navigation Satellite System (GNSS). For algorithm evaluation, this dataset releases ground truth of both localization and mapping, which are acquired by on-board Real-time Kinematic (RTK) and DJI L1 (post-processed by its supporting software DJI Terra), respectively. The dataset can be downloaded from: https://mars.hku.hk/dataset.html.
Persistent Identifierhttp://hdl.handle.net/10722/367886
ISSN
2023 Impact Factor: 7.5
2023 SCImago Journal Rankings: 4.346

 

DC FieldValueLanguage
dc.contributor.authorLi, Haotian-
dc.contributor.authorZou, Yuying-
dc.contributor.authorChen, Nan-
dc.contributor.authorLin, Jiarong-
dc.contributor.authorLiu, Xiyuan-
dc.contributor.authorXu, Wei-
dc.contributor.authorZheng, Chunran-
dc.contributor.authorLi, Rundong-
dc.contributor.authorHe, Dongjiao-
dc.contributor.authorKong, Fanze-
dc.contributor.authorCai, Yixi-
dc.contributor.authorLiu, Zheng-
dc.contributor.authorZhou, Shunbo-
dc.contributor.authorXue, Kaiwen-
dc.contributor.authorZhang, Fu-
dc.date.accessioned2025-12-19T08:00:10Z-
dc.date.available2025-12-19T08:00:10Z-
dc.date.issued2024-
dc.identifier.citationInternational Journal of Robotics Research, 2024, v. 43, n. 8, p. 1114-1127-
dc.identifier.issn0278-3649-
dc.identifier.urihttp://hdl.handle.net/10722/367886-
dc.description.abstractIn recent years, advancements in Light Detection and Ranging (LiDAR) technology have made 3D LiDAR sensors more compact, lightweight, and affordable. This progress has spurred interest in integrating LiDAR with sensors such as Inertial Measurement Units (IMUs) and cameras for Simultaneous Localization and Mapping (SLAM) research. Public datasets covering different scenarios, platforms, and viewpoints are crucial for multi-sensor fusion SLAM studies, yet most focus on handheld or vehicle-mounted devices with front or 360-degree views. Data from aerial vehicles with downward-looking views is scarce, existing relevant datasets usually feature low altitudes and are mostly limited to small campus environments. To fill this gap, we introduce the Multi-sensor Aerial Robots SLAM dataset (MARS-LVIG dataset), providing unique aerial downward-looking LiDAR-Visual-Inertial-GNSS data with viewpoints from altitudes between 80 m and 130 m. The dataset not only offers new aspects to test and evaluate existing SLAM algorithms, but also brings new challenges which can facilitate researches and developments of more advanced SLAM algorithms. The MARS-LVIG dataset contains 21 sequences, acquired across diversified large-area environments including an aero-model airfield, an island, a rural town, and a valley. Within these sequences, the UAV has speeds varying from 3 m/s to 12 m/s, a scanning area reaching up to 577,000 m<sup>2</sup>, and the max path length of 7.148 km in a single flight. This dataset encapsulates data collected by a lightweight, hardware-synchronized sensor package that includes a solid-state 3D LiDAR, a global-shutter RGB camera, IMUs, and a raw message receiver of the Global Navigation Satellite System (GNSS). For algorithm evaluation, this dataset releases ground truth of both localization and mapping, which are acquired by on-board Real-time Kinematic (RTK) and DJI L1 (post-processed by its supporting software DJI Terra), respectively. The dataset can be downloaded from: https://mars.hku.hk/dataset.html.-
dc.languageeng-
dc.relation.ispartofInternational Journal of Robotics Research-
dc.subjectaerial robots-
dc.subjectcamera-
dc.subjectDataset-
dc.subjectGlobal Navigation Satellite System-
dc.subjectInertial Measurement Unit-
dc.subjectLiDAR-
dc.subjectmulti-sensor fusion-
dc.subjectSimultaneous Localization and Mapping-
dc.titleMARS-LVIG dataset: A multi-sensor aerial robots SLAM dataset for LiDAR-visual-inertial-GNSS fusion-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1177/02783649241227968-
dc.identifier.scopuseid_2-s2.0-85183887891-
dc.identifier.volume43-
dc.identifier.issue8-
dc.identifier.spage1114-
dc.identifier.epage1127-
dc.identifier.eissn1741-3176-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats