Skip to content

BotanicGarden: A high-quality dataset for robot navigation in unstructured natural environments

Notifications You must be signed in to change notification settings

robot-pesg/BotanicGarden

Repository files navigation

BotanicGarden Dataset

A high-quality dataset for robot navigation in unstructured natural environments

Authors:

Yuanzhi Liu, Yujia Fu†, Minghui Qin†, Yufeng Xu†, Baoxin Xu, et al. († Contributed equally)

photo photo

Abstract:

The rapid developments of mobile robotics and autonomous navigation over the years are largely empowered by public datasets for testing and upgrading, such as sensor odometry and SLAM tasks. Impressive demos and benchmark scores have arisen, which may suggest the maturity of existing navigation techniques. However, these results are primarily based on moderate structured scenario testing. When transitioning to challenging unstructured environments, especially in GNSS-denied, texture-monotonous, and dense-vegetated natural fields, their performance can hardly sustain at a high level and requires further validation and improvement. To bridge this gap, we build a novel robot navigation dataset in a luxuriant botanic garden of more than 48000m2. Comprehensive sensors are used, including Gray and RGB stereo cameras, spinning and MEMS 3D LiDARs, and low-cost and industrial-grade IMUs, all of which are well calibrated and hardware-synchronized. An all-terrain wheeled robot is employed for data collection, traversing through thick woods, riversides, narrow trails, bridges, and grasslands, which are scarce in previous resources. This yields 33 short and long sequences, forming 17.1km trajectories in total. Excitedly, both highly-accurate ego-motions and 3D map ground truth are provided, along with fine-annotated vision semantics. We firmly believe that our dataset can advance robot navigation and sensor fusion research to a higher level.

photo photo photo photo photo photo photo photo photo photo photo photo photo photo photo

Main Contributions:

  • We build a novel multi-sensory dataset in an over 48000m2 botanic garden with 33 long and short sequences and 17.1km trajectories in total, containing dense and diverse natural elements that are scarce in previous resources.
  • We employed comprehensive sensors, including high-res and high-rate stereo gray and RGB cameras, spinning and MEMS 3D LiDARs, and low-cost and industrial-grade IMUs, supporting a wide range of applications. By elaborate development of the system, we have achieved highly-precise hardware-synchronization. Both the sensors availability and sync-quality are at top-level of this field.
  • We provide both highly-precise 3D map and trajectories ground truth by dedicated surveying works and advanced map-based localization algorithm. We also provide dense vision semantics labeled by experienced annotators. This is the first field robot navigation dataset that provides such all-sided and high-quality reference data.
video

Sensor Setup

photo
Sensor/Device Model Specification
Gray Stereo DALSA M1930 1920*1200, 2/3", 71°×56°FoV, 40Hz
RGB Stereo DALSA C1930 1920*1200, 2/3", 71°×56°FoV, 40Hz
LiDAR Velodyne VLP16 16C, 360°×30°FoV, ±3cm@100m, 10Hz
MEMS LiDAR Livox AVIA 70°×77°FoV, ±2cm@200m, 10Hz
D-GNSS/INS Xsens Mti-680G 9-axis, 400Hz, GNSS not in use
Consumer IMU BMI088 6-axis, 200Hz, Livox built-in
Wheel Encoder Scout V1.0 4WD, 3-axis, 200Hz
GT 3D Scanner Leica RTC360 130m range, 1mm+10ppm accuracy

Time Synchronization

photo
In a precise robot system with rich sensors and multi-hosts, time synchronization is extremely vital to eliminate perception delay and ensure navigation accuracy. Towards a high-quality dataset, we have taken very special cares on this problem. Our synchronization is based on a self-designed hardware trigger&timing board and a PTP-based network, as illustrated in the topological graph. The trigger and timing board is implemented by a compact MCU. It is programmed to produce three channels of pulses 1Hz-40Hz-400Hz in the very same phases. The 1Hz channel (pulse per second, PPS) is used for the synchronization of VLP-16 and AVIA accompanied with GPRMC signals; The 40Hz signal is used to trigger the cameras; And the 400Hz signal is used for triggering the Xsens IMU. The UTC time is maintained by MCU based on its onboard oscillator. Note that, to maintain the timing smoothness, we will never interrupt the MCU clock during the collections, instead, an UTC stamp will be conferred at the begin of each course-day via NTP or GNSS timing. So far, the LiDAR-camera-IMU chain has been completely synchronized in hardware, which can achieve sub-us precision as evaluated by the rising edge offset.
photo

Ground Truth Map

To ensure the global accuracy, we have not used any mobile-mapping based techniques (e.g., SLAM), instead we employ a tactical-grade stationary 3D laser scanner and conduct a qualified surveying and mapping job with professional colleagues from the College of Surveying and Geo-Informatics, Tongji University. The scanner is the RTC360 from Leica, which can output very dense and colored point cloud with a 130m scan radius and mm-level ranging accuracy, as shown the specifications in above table. The survey job takes in total 20 workdays and more than 900 individual scans, and get an accuracy of 11mm std. from Leica's report.

photo photo
Some survey photos and registration works:
photo photo

Data Sequences

Our dataset consists of 33 data sequences in total. The sequences info is shown below, including statistics of LiDAR scans, trajectory length, duration, images, and intersactions. Their trajectory images can be found here.

photo

At present, we have comprehensively evaluated the state-of-the-arts(SOTA) on 7 sample sequences. You can directly download their rosbags and raw files for benchmarking purposes. More sequences can be requested from Yuanzhi Liu via E-mail.

Stat/Sequence 1005-00 1005-01 1005-07 1006-01 1008-03 1018-00 1018-13
Size/GB1 66.8 49.0 59.8 83.1 71.0 13.0 20.9
rosbag2 onedrive onedrive onedrive onedrive onedrive onedrive onedrive
rosbag baidu baidu baidu baidu baidu baidu baidu
imagezip onedrive onedrive onedrive onedrive onedrive onedrive onedrive
GT-pose onedrive onedrive onedrive onedrive onedrive onedrive onedrive
Stat/Sequence 1005-05 1006-03 1008-01
Size/GB 31.4/6.2 29.6/5.9 34.4/6.7
VLIO-rosbag onedrive baidu onedrive baidu onedrive baidu
LIO-rosbag onedrive baidu onedrive baidu onedrive baidu
GT-pose onedrive baidu onedrive baidu onedrive baidu

The rostopics and corresponding message types are listed below:

ROS Topic Message Type Description
/dalsa_rgb/left/image_raw sensor_msgs/Image Left RGB camera
/dalsa_rgb/right/image_raw sensor_msgs/Image Right RGB camera
/dalsa_gray/left/image_raw sensor_msgs/Image Left Gray camera
/dalsa_gray/right/image_raw sensor_msgs/Image Right Gray camera
/velodyne_points sensor_msgs/PointCloud2 Velodyne VLP16 LiDAR
/livox/lidar livox_ros_driver/CustomMsg Livox AVIA LiDAR
/imu/data sensor_msgs/Imu Xsens IMU
/livox/imu3 sensor_msgs/Imu Livox BMI088 IMU
/gt_poses geometry_msgs/PoseStamped Ground truth poses

1Imagezip and no-vision rosbag size. 2The rosbags contain downsampled vision data (960x600@10Hz) to ease the downloads. Full res&rate frames (1920x1200@40Hz) are available in raw imagezips. 3The accelerometer of the BMI088 IMU has a gravity scale of 1. To use this IMU, multiply the accelerometer data by 9.8 to obtain acceleration in meters per second squared (m/s²). However, some algorithms designed specifically for Livox LiDAR may not require this multiplication (e.g., Fast-LIO2 and R3LIVE).

Ground Truth Trajectories

Our ground truth trajectories were generated within a survey-grade 3D map. With our dedicated survey works, the map was with ~1cm precision in the global coordinates, ensuring a cm-level precision for robot localization. All the 33 trajectories can be found in GT_traj folder.

photo photo

State-Of-The-Art Evaluation

We have tested the performance of visual (ORB-SLAM3), visual-inertial (ORB-SLAM3, VINS-Mono), LiDAR (LOAM), LiDAR-inertial (Fast-LIO2), and visual-LiDAR-inertial fusion (LVI-SAM, R3LIVE) systems on the above 7 sample sequences, check our Leaderboard for detail results.

Config Files

To simplify the user testing procedure, We have provided the calibration and config files of the State-Of-The-Arts, which can be accessed in calib and config folders.

Testing of LVI-SAM on 1005-00 sequence:

video

Semantic Dense Annotations

All data are provided in LabelMe format and support future reproducing. It is expected that these data can well facilitate robust motion estimation and semantic mapping. The number of labeled images, as well as the semantic classes and their distribution are shown below.

Sequence 1005-05 1005-07 1006-01 1006-03 1008-01
Labeled Images 233 206 270 216 256
photo
video video

Toolbox

Rosbag Conversion

Our dataset is captured in rosbag and raw formats. For the convenience of usage, we have provided a toolbox to convert between different structures, check the rosbag_tools folder for usage.

Semantic Conversion

The semantics are labelled in LabelMe json format. For the convenience of usage, we have provided a toolbox to convert to PASCAL VOC and MS COCO formats, check the semantic_tools folder for usage.

Calibration Tool

We have designed a concise toolbox for camera-LiDAR calibration based on several 2D checker boards. Check the calibration_tools folder for usage.

Evaluation

We recommend to use the open-source tool EVO for algorithm evaluation. Our Ground truth Poses are provided in TUM format consisting of timestamps, translations x-y-z, and quaternions x-y-z-w, which are concise and enable trajactory alignment based on time correspondances. Note that, the GT poses are tracking the VLP16 frame, so you must transform your poses to VLP16 side by hand-eye formula AX=XB before evaluation.

Acknowledgement

The authors would like to thank the colleagues from TongJi University and Sun Yat-sen University for their assistances in the rigorous survey works and post-processings, especially Xiaohang Shao, Chen Chen, and Kunhua Liu. We also thank A/Prof. Hangbin Wu for his guidance in data collection. Besides, we acknowledge Grace Xu from Livox for the support on Avia LiDAR, we acknowledge Claude Ng from Leica for the support on high-definition surveying, and we appreciate the colleagues of Appen for their professional works in visual semantic annotations. Yuanzhi Liu would like to thank Chenbo Gong for scene preparation work, and thank Jingxin Dong for her job-loggings and photographs during our data collection.

Funding

This work was supported by National Key R&D Program of China under Grant 2018YFB1305005.

Timeline

Feb  6, 2023   Open the GitHub website: https://github.com/robot-pesg/BotanicGarden
Feb 10, 2023   Rosbag tools open-sourced
Feb 13, 2023   Semantic tools open-sourced
Feb 14, 2023   Calibration params available
Feb 17, 2023   Calibration tools open-sourced
Feb 18, 2023   Paper submitted to IEEE Robotics and Automation Letters (RA-L)
Sep 18, 2023   Trajectories info updated
Sep 20, 2023   More sequences available
Dec 31, 2023   Update Leaderboard and invite more algorithms to join

Contact

This dataset is provided for academic purposes. If you meet some technical problems, please open an issue or contact <Yuanzhi Liu: [email protected]> and <Yufeng Xu: [email protected]>.

Citation

If our dataset helps your research, please cite:

@ARTICLE{liu2023botanicgarden,
  author={Liu, Yuanzhi and Fu, Yujia and Qin, Minghui and Xu, Yufeng and Xu, Baoxin and Chen, Fengdong and Goossens, Bart and Sun, Poly Z.H. and Yu, Hongwei and Liu, Chun and Chen, Long and Tao, Wei and Zhao, Hui},
  journal={IEEE Robotics and Automation Letters}, 
  title={BotanicGarden: A High-Quality Dataset for Robot Navigation in Unstructured Natural Environments}, 
  year={2024},
  volume={9},
  number={3},
  pages={2798-2805},
  doi={10.1109/LRA.2024.3359548}}

About

BotanicGarden: A high-quality dataset for robot navigation in unstructured natural environments

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published