Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
144 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GRIL-Calib: Targetless Ground Robot IMU-LiDAR Extrinsic Calibration Method using Ground Plane Motion Constraints (2312.14035v2)

Published 21 Dec 2023 in cs.RO

Abstract: Targetless IMU-LiDAR extrinsic calibration methods are gaining significant attention as the importance of the IMU-LiDAR fusion system increases. Notably, existing calibration methods derive calibration parameters under the assumption that the methods require full motion in all axes. When IMU and LiDAR are mounted on a ground robot the motion of which is restricted to planar motion, existing calibration methods are likely to exhibit degraded performance. To address this issue, we present GRIL-Calib: a novel targetless Ground Robot IMU-LiDAR Calibration method. Our proposed method leverages ground information to compensate for the lack of unrestricted full motion. First, we propose LiDAR Odometry (LO) using ground plane residuals to enhance calibration accuracy. Second, we propose the Ground Plane Motion (GPM) constraint and incorporate it into the optimization for calibration, enabling the determination of full 6-DoF extrinsic parameters, including theoretically unobservable direction. Finally, unlike baseline methods, we formulate the calibration not as sequential two optimizations but as a single optimization (SO) problem, solving all calibration parameters simultaneously and improving accuracy. We validate our GRIL-Calib by applying it to various real-world datasets and comparing its performance with that of existing state-of-the-art methods in terms of accuracy and robustness. Our code is available at https://github.com/Taeyoung96/GRIL-Calib.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. I. Vizzo, T. Guadagnino, B. Mersch, L. Wiesmann, J. Behley, and C. Stachniss, “Kiss-icp: In defense of point-to-point icp–simple, accurate, and robust registration if done the right way,” IEEE Robotics and Automation Letters, vol. 8, no. 2, pp. 1029–1036, 2023.
  2. J. Wang, M. Xu, G. Zhao, and Z. Chen, “3d lidar localization based on novel nonlinear optimization method for autonomous ground robot,” IEEE Transactions on Industrial Electronics, 2023.
  3. J. Zhang and S. Singh, “Loam: Lidar odometry and mapping in real-time.” in Robotics: Science and systems, vol. 2, no. 9.   Berkeley, CA, 2014, pp. 1–9.
  4. W. Xu, Y. Cai, D. He, J. Lin, and F. Zhang, “Fast-lio2: Fast direct lidar-inertial odometry,” IEEE Transactions on Robotics, vol. 38, no. 4, pp. 2053–2073, 2022.
  5. T. Shan, B. Englot, D. Meyers, W. Wang, C. Ratti, and D. Rus, “Lio-sam: Tightly-coupled lidar inertial odometry via smoothing and mapping,” in 2020 IEEE/RSJ international conference on intelligent robots and systems (IROS).   IEEE, 2020, pp. 5135–5142.
  6. C. Bai, T. Xiao, Y. Chen, H. Wang, F. Zhang, and X. Gao, “Faster-lio: Lightweight tightly coupled lidar-inertial odometry using parallel sparse incremental voxels,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 4861–4868, 2022.
  7. S. Li, X. Li, Y. Zhou, and C. Xia, “Targetless extrinsic calibration of lidar–imu system using raw gnss observations for vehicle applications,” IEEE Transactions on Instrumentation and Measurement, vol. 72, pp. 1–11, 2023.
  8. W. Liu and Y. Li, “Error modeling and extrinsic–intrinsic calibration for lidar-imu system based on cone-cylinder features,” Robotics and Autonomous Systems, vol. 114, pp. 124–133, 2019.
  9. Y. Yang, P. Geneva, K. Eckenhoff, and G. Huang, “Degenerate motion analysis for aided ins with online spatial and temporal sensor calibration,” IEEE Robotics and Automation Letters, vol. 4, no. 2, pp. 2070–2077, 2019.
  10. Y. Yang, P. Geneva, X. Zuo, and G. Huang, “Online self-calibration for visual-inertial navigation: Models, analysis, and degeneracy,” IEEE Transactions on Robotics, 2023.
  11. J. Lv, X. Zuo, K. Hu, J. Xu, G. Huang, and Y. Liu, “Observability-aware intrinsic and extrinsic calibration of lidar-imu systems,” IEEE Transactions on Robotics, vol. 38, no. 6, pp. 3734–3753, 2022.
  12. X. Wei, J. Lv, J. Sun, E. Dong, and S. Pu, “Gclo: Ground constrained lidar odometry with low-drifts for gps-denied indoor environments,” in 2022 International Conference on Robotics and Automation (ICRA).   IEEE, 2022, pp. 2229–2235.
  13. H. Li and J. Stueckler, “Visual-inertial odometry with online calibration of velocity-control based kinematic motion models,” IEEE Robotics and Automation Letters, vol. 7, no. 3, pp. 6415–6422, 2022.
  14. J. Yin, A. Li, T. Li, W. Yu, and D. Zou, “M2dgr: A multi-sensor and multi-scenario slam dataset for ground robots,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 2266–2273, 2021.
  15. M. Helmberger, K. Morin, B. Berner, N. Kumar, G. Cioffi, and D. Scaramuzza, “The hilti slam challenge dataset,” IEEE Robotics and Automation Letters, vol. 7, no. 3, pp. 7518–7525, 2022.
  16. D. Feng, Y. Qi, S. Zhong, Z. Chen, Y. Jiao, Q. Chen, T. Jiang, and H. Chen, “S3e: A large-scale multimodal dataset for collaborative slam,” arXiv preprint arXiv:2210.13723, 2022.
  17. C. Le Gentil, T. Vidal-Calleja, and S. Huang, “3d lidar-imu calibration based on upsampled preintegrated measurements for motion distortion correction,” in 2018 IEEE International Conference on Robotics and Automation (ICRA), 2018, pp. 2149–2155.
  18. S. Mishra, G. Pandey, and S. Saripalli, “Target-free extrinsic calibration of a 3d-lidar and an imu,” in 2021 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI).   IEEE, 2021, pp. 1–7.
  19. J. Lv, J. Xu, K. Hu, Y. Liu, and X. Zuo, “Targetless calibration of lidar-imu system based on continuous-time batch estimation,” in 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2020, pp. 9968–9975.
  20. F. Zhu, Y. Ren, and F. Zhang, “Robust real-time lidar-inertial initialization,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2022, pp. 3948–3955.
  21. S. Madgwick et al., “An efficient orientation filter for inertial and inertial/magnetic sensor arrays,” Report x-io and University of Bristol (UK), vol. 25, pp. 113–118, 2010.
  22. S. Lee, H. Lim, and H. Myung, “Patchwork++: Fast and robust ground segmentation solving partial under-segmentation using 3d point cloud,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2022, pp. 13 276–13 283.
  23. C. Hertzberg, R. Wagner, U. Frese, and L. Schröder, “Integrating generic sensor fusion algorithms with sound state representations through encapsulation of manifolds,” Information Fusion, vol. 14, no. 1, pp. 57–77, 2013.
  24. D. Baleanu, O. Defterli, and O. P. Agrawal, “A central difference numerical scheme for fractional optimal control problems,” Journal of Vibration and Control, vol. 15, no. 4, pp. 583–597, 2009.
  25. F. Gustafsson, “Determining the initial states in forward-backward filtering,” IEEE Transactions on signal processing, vol. 44, no. 4, pp. 988–992, 1996.
  26. W. Xu, D. He, Y. Cai, and F. Zhang, “Robots’ state estimation and observability analysis based on statistical motion models,” IEEE Transactions on Control Systems Technology, vol. 30, no. 5, pp. 2030–2045, 2022.
  27. S. Xu, J. S. Willners, Z. Hong, K. Zhang, Y. R. Petillot, and S. Wang, “Observability-aware active extrinsic calibration of multiple sensors,” in 2023 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2023, pp. 2091–2097.
Citations (1)

Summary

  • The paper introduces GRIL-Calib, a novel method that integrates ground plane constraints for simultaneous 6-DoF IMU-LiDAR extrinsic calibration without requiring predefined targets.
  • It employs a unified optimization approach that mitigates z-axis drift and improves odometry accuracy in environments with inherent planar motion.
  • Validation on multiple datasets shows significantly reduced rotation and translation RMSE compared to existing methods, demonstrating robustness in real-world robotic applications.

Overview of GRIL-Calib: Targetless Ground Robot IMU-LiDAR Extrinsic Calibration

The paper "GRIL-Calib: Targetless Ground Robot IMU-LiDAR Extrinsic Calibration Method using Ground Plane Motion Constraints" addresses the critical problem of calibrating the spatial transformations between Inertial Measurement Units (IMUs) and Light Detection and Ranging (LiDAR) systems on ground robots. This task is pivotal for the effective fusion of IMU-LiDAR data, which is widely used in robotics for applications such as localization, mapping, and motion estimation. Traditional calibration methods often necessitate full motion along all axes, which is not feasible for ground robots due to their inherent planar motion constraints. This paper proposes GRIL-Calib, a novel calibration framework that leverages ground plane motion constraints to estimate accurate 6-DoF extrinsic parameters without relying on predefined targets or extensive motion.

Key Contributions and Methodology

GRIL-Calib innovatively integrates ground plane information to compensate for the lack of full-motion observability, addressing a significant gap in existing calibration methodologies. It comprises three primary contributions:

  1. LiDAR Odometry with Ground Plane Residuals: The method enhances LiDAR odometry by integrating ground plane constraints, which mitigates the drift typically observed around the z-axis when operating in planar motion. This involves introducing a ground plane residual into an Iterative Error State Kalman Filter (IESKF), thus improving the odometry accuracy and stability.
  2. Ground Plane Motion (GPM) Constraints: Unlike baseline methods which treat the estimation of extrinsic parameters sequentially, GRIL-Calib employs GPM constraints in a single optimization framework. This allows the simultaneous estimation of the entire parameter set, including directions that are theoretically unobservable in planar motion. The GPM constraints effectively exploit geometric relationships between the sensors and the ground, ensuring robust calibration even with restricted motions.
  3. Single Optimization Approach: This approach resolves the calibration as a unified optimization problem instead of splitting it into sequential tasks for translation and rotation. This method benefits from the correlations between different parameter sets, significantly enhancing the calibration accuracy.

Validation and Implications

The authors validate GRIL-Calib on three publicly available datasets: M2DGR, Hilti, and S3E, showing that it consistently outperforms existing state-of-the-art targetless methods. It particularly excels in reducing root mean square error (RMSE) in both rotation and translation parameters across various scenarios and robot platforms, evidencing its robustness and adaptability.

The theoretical implications of this research are profound, as it expands the capabilities of sensor fusion in constrained motion environments by achieving accurate calibration in situations previously deemed challenging or impossible. Practically, the robust performance of GRIL-Calib in diverse datasets suggests its potential utility in real-world robotic systems, where accurate motion estimation is critical for autonomous operations, especially in outdoor and cluttered environments.

Future Directions

The paper opens several pathways for subsequent research. Future work may explore extending the framework to accommodate scenarios with uneven or dynamic ground surfaces, a natural extension given the method's reliance on ground plane constraints. Moreover, integrating additional sensor modalities and investigating adaptive weighting strategies in the optimization could further enhance calibration robustness. Additionally, as autonomous systems continue to evolve, adapting these calibration strategies to account for changes in sensor performance over time will be crucial for maintaining long-term accuracy.

In conclusion, GRIL-Calib represents a significant advance in the calibration of IMU-LiDAR systems for ground robots, providing a reliable solution to a previously challenging problem. Its emphasis on exploiting available ground plane information within a robust optimization framework positions it as a valuable tool in the continued advancement of robotic navigation and perception systems.