Ground-VIO: Monocular Visual-Inertial Odometry with Online Calibration of Camera-Ground Geometric Parameters (2306.08341v2)
Abstract: Monocular visual-inertial odometry (VIO) is a low-cost solution to provide high-accuracy, low-drifting pose estimation. However, it has been meeting challenges in vehicular scenarios due to limited dynamics and lack of stable features. In this paper, we propose Ground-VIO, which utilizes ground features and the specific camera-ground geometry to enhance monocular VIO performance in realistic road environments. In the method, the camera-ground geometry is modeled with vehicle-centered parameters and integrated into an optimization-based VIO framework. These parameters could be calibrated online and simultaneously improve the odometry accuracy by providing stable scale-awareness. Besides, a specially designed visual front-end is developed to stably extract and track ground features via the inverse perspective mapping (IPM) technique. Both simulation tests and real-world experiments are conducted to verify the effectiveness of the proposed method. The results show that our implementation could dramatically improve monocular VIO accuracy in vehicular scenarios, achieving comparable or even better performance than state-of-art stereo VIO solutions. The system could also be used for the auto-calibration of IPM which is widely used in vehicle perception. A toolkit for ground feature processing, together with the experimental datasets, would be made open-source (https://github.com/GREAT-WHU/gv_tools).
- A. Geiger, P. Lenz and R. Urtasun, “Are we ready for autonomous driving? The KITTI vision benchmark suite,” 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 2012, pp. 3354-3361.
- X. Chen, K. Kundu, Z. Zhang, H. Ma, S. Fidler and R. Urtasun, “Monocular 3D Object Detection for Autonomous Driving,” 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 2016, pp. 2147-2156.
- H. Lategahn, A. Geiger and B. Kitt, “Visual SLAM for autonomous ground vehicles,” 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 2011, pp. 1732-1737.
- G. Huang, “Visual-Inertial Navigation: A Concise Review,” 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 2019, pp. 9572-9582.
- T. Qin, P. Li and S. Shen, “VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator,” in IEEE Transactions on Robotics, vol. 34, no. 4, pp. 1004-1020, Aug. 2018.
- A. Martinelli, “Visual-inertial structure from motion: Observability and resolvability,” in Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst., pp. 4235-4242, Nov. 2013.
- J. Hernandez, K. Tsotsos and S. Soatto, “Observability identifiability and sensitivity of vision-aided inertial navigation,” in Proc. IEEE Int. Conf. Robot. Autom. (ICRA), pp. 2319-2325, May 2015.
- Y. Yang and G. Huang, “Observability Analysis of Aided INS With Heterogeneous Features of Points, Lines, and Planes,” in IEEE Transactions on Robotics, vol. 35, no. 6, pp. 1399-1418, Dec. 2019.
- W. Lee, Y. Yang and G. Huang, “Efficient Multi-sensor Aided Inertial Navigation with Online Calibration,” 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 2021, pp. 5706-5712.
- S. Li, X. Li, H. Wang, Y. Zhou and Z. Shen, “Multi-GNSS PPP/INS/Vision/LiDAR tightly integrated system for precise navigation in urban environments,” Information Fusion, vol. 90, 2023, pp. 218-232.
- J. Jeong and A. Kim, “Adaptive Inverse Perspective Mapping for lane map generation with SLAM,” 2016 13th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Xi’an, China, 2016, pp. 38-41.
- J. Wang, T. Mei, B. Kong and H. Wei, “An approach of lane detection based on Inverse Perspective Mapping,” 17th International IEEE Conference on Intelligent Transportation Systems (ITSC), Qingdao, 2014, pp. 35-38.
- Y. -L. Chang, L. -Y. Hsu and O. T. . -C. Chen, “Auto-Calibration Around-View Monitoring System,” 2013 IEEE 78th Vehicular Technology Conference (VTC Fall), Las Vegas, NV, USA, 2013, pp. 1-5.
- J. Jeong, Y. Cho and A. Kim, “Road-SLAM : Road marking based SLAM with lane-level accuracy,” 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 2017, pp. 1736-1473.
- Y. Zhou, X. Li, S. Li and X. Wang, “Visual Mapping and Localization System Based on Compact Instance-Level Road Markings With Spatial Uncertainty,” in IEEE Robotics and Automation Letters, vol. 7, no. 4, pp. 10802-10809, Oct. 2022.
- N. Gosala and A. Valada, “Bird’s-Eye-View Panoptic Segmentation Using Monocular Frontal View Images,” in IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 1968-1975, April 2022.
- A. I. Mourikis and S. I. Roumeliotis, “A Multi-State Constraint Kalman Filter for Vision-aided Inertial Navigation,” Proceedings 2007 IEEE International Conference on Robotics and Automation, Rome, Italy, 2007, pp. 3565-3572.
- M. Li and A. I. Mourikis, “Improving the accuracy of EKF-based visual-inertial odometry,” 2012 IEEE International Conference on Robotics and Automation, Saint Paul, MN, USA, 2012, pp. 828-835.
- K. Eckenhoff, P. Geneva and G. Huang, “MIMC-VINS: A Versatile and Resilient Multi-IMU Multi-Camera Visual-Inertial Navigation System,” in IEEE Transactions on Robotics, vol. 37, no. 5, pp. 1360-1380, Oct. 2021.
- P. Geneva, K. Eckenhoff, W. Lee, Y. Yang and G. Huang, “OpenVINS: A Research Platform for Visual-Inertial Estimation,” 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 2020, pp. 4666-4672.
- C. Forster, L. Carlone, F. Dellaert and D. Scaramuzza, “On-Manifold Preintegration for Real-Time Visual–Inertial Odometry,” in IEEE Transactions on Robotics, vol. 33, no. 1, pp. 1-21, Feb. 2017.
- S. Leutenegger, S. Lynen, M. Bosse, R. Siegwart, P. Furgale, “Keyframe-based visual-inertial odometry using nonlinear optimization”, International Journal of Robotics Research (IJRR), 2014.
- Tong Qin, Jie Pan, Shaozu Cao, Shaojie Shen, “A General Optimization-based Framework for Local Odometry Estimation with Multiple Sensors”, arXiv:1901.03638 [cs.CV], Jan. 2019.
- C. Campos, R. Elvira, J. J. G. Rodríguez, J. M. M. Montiel and J. D. Tardós, “ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual-Inertial, and Multimap SLAM,” in IEEE Transactions on Robotics, vol. 37, no. 6, pp. 1874-1890, Dec. 2021.
- P. Zhou, Y. Liu, P. Gu, J. Liu and Z. Meng, “Visual Localization and Mapping Leveraging the Constraints of Local Ground Manifolds,” in IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 4196-4203, April 2022.
- K. Konolige, G. Grisetti, R. Kümmerle, W. Burgard, B. Limketkai and R. Vincent, “Efficient Sparse Pose Adjustment for 2D mapping,” 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 2010, pp. 22-29.
- D. Scaramuzza, F. Fraundorfer and R. Siegwart, “Real-time monocular visual odometry for on-road vehicles with 1-point RANSAC,” 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 2009, pp. 4293-4299.
- F. Zheng and Y. -H. Liu, “SE(2)-Constrained Visual Inertial Fusion for Ground Vehicles,” in IEEE Sensors Journal, vol. 18, no. 23, pp. 9699-9707, 1 Dec.1, 2018.
- F. Zheng and Y. -H. Liu, “Visual-Odometric Localization and Mapping for Ground Vehicles Using SE(2)-XYZ Constraints,” 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 2019, pp. 3556-3562.
- R. Kang, L. Xiong, M. Xu, J. Zhao and P. Zhang, “VINS-Vehicle: A Tightly-Coupled Vehicle Dynamics Extension to Visual-Inertial State Estimator,” 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 2019, pp. 3593-3600.
- M. Zhang, X. Zuo, Y. Chen, Y. Liu and M. Li, “Pose Estimation for Ground Robots: On Manifold Representation, Integration, Reparameterization, and Optimization,” in IEEE Transactions on Robotics, vol. 37, no. 4, pp. 1081-1099, Aug. 2021.
- M. Ouyang, Z. Cao, P. Guan, Z. Li, C. Zhou and J. Yu, “Visual-Gyroscope-Wheel Odometry With Ground Plane Constraint for Indoor Robots in Dynamic Environment,” in IEEE Sensors Letters, vol. 5, no. 3, pp. 1-4, March 2021, Art no. 6000504.
- S. Song, M. Chandraker and C. C. Guest, “High Accuracy Monocular SFM and Scale Correction for Autonomous Driving,” in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no. 4, pp. 730-743, 1 April 2016.
- B. Lee, K. Daniilidis and D. D. Lee, “Online self-supervised monocular visual odometry for ground vehicles,” 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 2015, pp. 5232-5238.
- R. Tian, Y. Zhang, D. Zhu, S. Liang, S. Coleman and D. Kerr, “Accurate and Robust Scale Recovery for Monocular Visual Odometry Based on Plane Geometry,” 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 2021, pp. 5296-5302.
- J. Shi and C. Tomasi, “Good features to track,” in Proc. IEEE Int. Conf. Pattern Recog., pp. 593-600, 1994.
- B. D. Lucas and T. Kanade, “An iterative image registration technique with an application to stereo vision,” in Proc. Int. Joint Conf. Artif. Intell., pp. 24-28, Aug. 1981.
- G. Bradski, “The OpenCV Library.” [Online]. Available: https://docs.opencv.org/3.4/d9/d0c/group__calib3d.html
- Liang-Chieh Chen, Yukun Zhu, George Papandreou, Florian Schroff, Hartwig Adam, “Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation”, in Proc. Eur. Conf. Comput. Vis., 2018, pp. 801-818
- S. Agarwal and K. Mierle, “Ceres solver.” [Online]. Available: http://ceres-solver.org
- A. Dosovitskiy, G. Ros, F. Codevilla, A. Lopez and V. Koltun, “CARLA: An Open Urban Driving Simulator,” Proceedings of the 1st Annual Conference on Robot Learning ser. Proceedings of Machine Learning Research, vol. 78, pp. 1-16, Nov. 2017.