VILO SLAM: Tightly Coupled Binocular Vision–Inertia SLAM Combined with LiDAR
<p>The system structure of VILO.</p> "> Figure 2
<p>Camera-lidar time alignment diagram.</p> "> Figure 3
<p>Binocular vision-IMU-2D lidar pose-estimation factor graph.</p> "> Figure 4
<p>Sensor mounting platforms. PC is used for data acquisition of robot sensors, pose estimation and mapping. Embedded controller is used for robot motion control.</p> "> Figure 5
<p>Indoor corridor environment. (<b>a</b>) Corridor outline; (<b>b</b>) Corner of the corridor; (<b>c</b>) Insufficient light inside the corridor; (<b>d</b>) Corridor hall scene.</p> "> Figure 6
<p>The trajectory of each algorithm (top view). (<b>a</b>) Standard trajectory; (<b>b</b>) This paper; (<b>c</b>) VINS-Fusion; (<b>d</b>) ORB-SLAM2.</p> "> Figure 7
<p>3D mapping comparison experiment based on three types of pose-estimation algorithms. (<b>a</b>) Based on ORB_SLAM2; (<b>b</b>) Based on VINS_Fusion; (<b>c</b>) Based on algorithm of this paper.</p> ">
Abstract
:1. Introduction
- Tightly coupling low-cost 2D lidar observations with stereo vision and inertial observations improves the accuracy and robustness of pose estimation in traditional visual–inertial SLAM algorithms in scenarios where visual features are lost due to darkness, strong light, or lack of texture.
- A lidar residual factor is constructed using the 2D lidar odometry model, and the Jacobian matrix of lidar residuals with respect to the state variables to be estimated is derived.
- The residual constraint equation of vision-IMU-LiDAR is constructed, and optimal robot pose estimation is obtained using nonlinear optimization, which solves the problem of fusing 2D lidar observations with binocular visual–inertial information in a tightly coupled manner.
2. Related Work
3. Multi-Sensor Pose Estimation Based on Tightly Coupled Optimization
3.1. Nonlinear Least-Squares Model Based on Vision-IMU-Lidar
3.1.1. Visual Reprojection Residual Constraint
3.1.2. IMU Residual Constraints
3.2. 2D Lidar Residual Error Constraints
3.2.1. 2D lidar Odometry Algorithm
3.2.2. Residual Items Based on 2D Lidar Odometry Model
3.3. Pseudo-Code Description
Algorithm 1 VILO Residual Calculation and Optimization | |
Input: IMU data, camera images, and 2D LiDAR point cloud data | |
Output: in the world coordinate system | |
// Step 1. Initialize system and variables | |
1: | //Initialize state variables according to Equation (1); |
2: | //Initialize observation set according to Equation (2); |
3: | // Step 2. Receive data and perform pose estimation and optimization |
4: | while received IMU, camera image and 2D LiDAR point cloud do |
5: | Undistorted pcd = Motion distortion correction (imu data, point cloud) |
6: | //Time alignment |
7: | |
8: | |
9: | end if |
10: | //IMU residual according to Equation (7) |
11: | //Extract features (pixel coordinates) and optical flow tracking |
12: | //Visual reprojection residual according to Equations (5) and (6) |
13: | //LiDAR pose estimation according to Equations (8)–(15) |
14: | //Construct 2D lidar odometry residual according to Equation (18) |
15: | //Ceres optimization, get output pose |
16: | end while |
4. Experiment
4.1. Motion Trajectory Comparison Experiment
- (1)
- Parts of the corridor walls are completely white without obvious visual features;
- (2)
- In the corridor hall, the walls are covered with tiles with high reflectivity, which affects the camera observation data;
- (3)
- The similarity of some corridor scenes is relatively high, and there are no special markings, which affects the accuracy of 2D lidar odometry;
- (4)
- On some floors of the corridor, there are cracks, uneven heights, and large vibrations when the robot moves, causing large fluctuations in the measured values of each sensor;
- (5)
- During the experiment, there were many pedestrians in the field of view of the camera and 2D lidar.
4.2. Pose Estimation Accuracy Verification Experiment
4.3. Comparative Experiment on Dense Mapping of Spatial Environment Based on Three Types of Pose-Estimation Algorithms
5. Summary
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Engel, J.; Schöps, T.; Cremers, D. LSD-SLAM: Large-scale direct monocular SLAM. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Part II 13, Zurich, Switzerland, 6–12 September 2014; Springer International Publishing: Berlin/Heidelberg, Germany, 2014. [Google Scholar]
- Engel, J.; Koltun, V.; Cremers, D. Direct sparse odometry. IEEE Transact. Pattern anal. Machine Intel. 2017, 40, 611–625. [Google Scholar] [CrossRef]
- Forster, C.; Pizzoli, M.; Scaramuzza, D. SVO: Fast semi-direct monocular visual odometry. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 15–22. [Google Scholar]
- Lahemer, E.S.; Rad, A. An adaptive augmented vision-based ellipsoidal SLAM for indoor environments. Sensors 2019, 19, 2795. [Google Scholar] [CrossRef] [PubMed]
- Lynen, S.; Achtelik, M.W.; Weiss, S.; Chli, M.; Siegwart, R. A robust and modular multi-sensor fusion approach applied to mav navigation. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 3923–3929. [Google Scholar]
- López, E.; García, S.; Barea, R.; Bergasa, L.M.; Molinos, E.J.; Arroyo, R.; Romera, E.; Pardo, S. A multi-sensorial simultaneous localization and mapping (SLAM) system for low-cost micro aerial vehicles in GPS-denied environments. Sensors 2017, 17, 802. [Google Scholar] [CrossRef] [PubMed]
- Yan, Y.; Zhang, B.; Zhou, J.; Zhang, Y.; Liu, X. Real-Time Localization and Mapping Utilizing Multi-Sensor Fusion and Visual–IMU–Wheel Odometry for Agricultural Robots in Unstructured, Dynamic and GPS-Denied Greenhouse Environments. Agronomy 2022, 12, 1740. [Google Scholar] [CrossRef]
- Leutenegger, S.; Lynen, S.; Bosse, M.; Siegwart, R.; Furgale, P. Keyframe-based visual–inertial odometry using nonlinear optimization. Int. J. Robot. Res. 2015, 34, 314–334. [Google Scholar] [CrossRef]
- Forster, C.; Carlone, L.; Dellaert, F.; Scaramuzzza, D. IMU preintegration on manifold for efficient visual-inertial maximum-a-posteriori estimation. In Robotics Science and Systems; Georgia Institute of Technology: Roma, Italy, 2015; pp. 1–20. [Google Scholar]
- Campos, C.; Elvira, R.; Rodríguez, J.J.G.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual-Inertial and Multi-Map SLAM. IEEE Transact. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
- Xiao, J.; Xiong, D.; Yu, Q.; Huang, K.; Lu, H.; Zeng, Z. A Real-Time Sliding-Window-Based Visual-Inertial Odometry for MAVs. IEEE Transact. Ind. Inform. 2020, 16, 4049–4058. [Google Scholar] [CrossRef]
- Wang, Y.; Li, Z.; Su, C.-Y. Multisensor-Based Navigation and Control of a Mobile Service Robot. IEEE Trans. Syst. Man Cybern. Syst. 2019, 49, 1–11. [Google Scholar]
- Hashim, H.A.; Eltoukhy, A.E.E. Nonlinear Filter for Simultaneous Localization and Mapping on a Matrix Lie Group Using IMU and Feature Measurements. IEEE Transact. Syst. Man Cybernet. Syst. 2022, 52, 2098–2109. [Google Scholar] [CrossRef]
- Qin, T.; Cao, S.; Pan, J.; Shen, S. A general optimization-based framework for global pose estimation with multiple sensors. arXiv 2019, arXiv:1901.03642. [Google Scholar]
- Tixiao, S.; Englot, B.; Meyers, D.; WAmg, W.; Ratti, C.; Rus, D. Lio-sam: Tightly-coupled lidar inertial odometry via smoothing and mapping. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 10 February 2020. [Google Scholar]
- Jiarong, L.; Zhang, F. R3LIVE: A Robust, Real-time, RGB-colored, LiDAR-Inertial-Visual tightly-coupled state Estimation and mapping package. In Proceedings of the 22 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022. [Google Scholar]
- Ming, J. Binocular Vision Positioning and 2D Lidar Mapping of Indoor Environment Based on Point-Line FEATURE. Master’s Thesis, Southwest Jiaotong University, Chengdu, China, 2018; p. 91. [Google Scholar]
- Qin, T.; Li, P.; Shen, S. VINS-Mono: A robust and versatile monocular visual-inertial state estimator. IEEE Trans. Robot 2018, 34, 1004–1020. [Google Scholar] [CrossRef]
- Hess, W.; Kohler, D.; Rapp, H.; Andor, D. Real-time loop closure in 2D LIDAR SLAM. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 9 June 2016. [Google Scholar]
- Wang, K.; Gao, F.; Shen, S. Real-time Scalable Dense Surfel Mapping. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 16–21 May 2019; pp. 6919–6925. [Google Scholar]
Algorithm | Parameters | Initial Pose | 1st Finish Pose | 2nd Finish Pose | 3rd Finish Pose | Average | MAE |
---|---|---|---|---|---|---|---|
ORB-SLAM2 | X axis offset (m) | 0.032 | 6.56 | 5.36 | 7.28 | 6.40 | 6.368 |
Y axis offset (m) | −0.028 | −5.32 | −5.98 | −6.46 | −5.92 | 5.892 | |
yaw angle (degrees) | 0.351 | 8.563 | 7.253 | 10.26 | 8.692 | 8.341 | |
VINS-Fusion | X axis offset (m) | −0.021 | 5.221 | 4.336 | 4.758 | 4.772 | 4.793 |
Y axis offset (m) | 0.033 | −4.532 | −5.142 | −4.349 | −4.674 | 4.707 | |
yaw angle (degrees) | 0.283 | 5.286 | 5.463 | 5.852 | 5.534 | 5.251 | |
Our method | X axis offset (m) | −0.011 | −3.326 | −3.635 | −3.867 | −3.609 | 3.598 |
Y axis offset (m) | 0.016 | −3.855 | −4.126 | −3.732 | −3.904 | 3.920 | |
yaw angle (degrees) | −0.203 | 4.563 | 4.068 | 3.659 | 4.10 | 4.303 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Peng, G.; Zhou, Y.; Hu, L.; Xiao, L.; Sun, Z.; Wu, Z.; Zhu, X. VILO SLAM: Tightly Coupled Binocular Vision–Inertia SLAM Combined with LiDAR. Sensors 2023, 23, 4588. https://doi.org/10.3390/s23104588
Peng G, Zhou Y, Hu L, Xiao L, Sun Z, Wu Z, Zhu X. VILO SLAM: Tightly Coupled Binocular Vision–Inertia SLAM Combined with LiDAR. Sensors. 2023; 23(10):4588. https://doi.org/10.3390/s23104588
Chicago/Turabian StylePeng, Gang, Yicheng Zhou, Lu Hu, Li Xiao, Zhigang Sun, Zhangang Wu, and Xukang Zhu. 2023. "VILO SLAM: Tightly Coupled Binocular Vision–Inertia SLAM Combined with LiDAR" Sensors 23, no. 10: 4588. https://doi.org/10.3390/s23104588