Design and Evaluation of a Generic Visual SLAM Framework for Multi-Camera Systems (2210.07315v2)
Abstract: Multi-camera systems have been shown to improve the accuracy and robustness of SLAM estimates, yet state-of-the-art SLAM systems predominantly support monocular or stereo setups. This paper presents a generic sparse visual SLAM framework capable of running on any number of cameras and in any arrangement. Our SLAM system uses the generalized camera model, which allows us to represent an arbitrary multi-camera system as a single imaging device. Additionally, it takes advantage of the overlapping fields of view (FoV) by extracting cross-matched features across cameras in the rig. This limits the linear rise in the number of features with the number of cameras and keeps the computational load in check while enabling an accurate representation of the scene. We evaluate our method in terms of accuracy, robustness, and run time on indoor and outdoor datasets that include challenging real-world scenarios such as narrow corridors, featureless spaces, and dynamic objects. We show that our system can adapt to different camera configurations and allows real-time execution for typical robotic applications. Finally, we benchmark the impact of the critical design parameters - the number of cameras and the overlap between their FoV that define the camera configuration for SLAM. All our software and datasets are freely available for further research.
- A. J. Davison, “Real-time simultaneous localisation and mapping with a single camera,” in Computer Vision, IEEE International Conference on, vol. 3. IEEE Computer Society, 2003, pp. 1403–1403.
- C. Campos, R. Elvira, J. J. G. Rodríguez, J. M. Montiel, and J. D. Tardós, “ORBSlam3: An Accurate Open-Source Library for Visual, Visual–Inertial, and Multimap SLAM,” IEEE Transactions on Robotics, 2021.
- R. A. Newcombe, S. J. Lovegrove, and A. J. Davison, “DTAM: Dense tracking and mapping in real-time,” in 2011 international conference on computer vision. IEEE, 2011, pp. 2320–2327.
- C. Forster, Z. Zhang, M. Gassner, M. Werlberger, and D. Scaramuzza, “SVO: Semidirect Visual Odometry for Monocular and Multicamera Systems,” IEEE Transactions on Robotics, vol. 33, pp. 249–265, 2017.
- M. He, C. Zhu, Q. Huang, B. Ren, and J. Liu, “A review of monocular visual odometry,” The Visual Computer, vol. 36, pp. 1053–1065, 2020.
- J. Engel, J. Stückler, and D. Cremers, “Large-scale direct SLAM with stereo cameras,” in 2015 IEEE/RSJ international conference on intelligent robots and systems (IROS). IEEE, 2015, pp. 1935–1942.
- C. Kerl, J. Sturm, and D. Cremers, “Robust odometry estimation for RGB-D cameras,” in 2013 IEEE international conference on robotics and automation. IEEE, 2013, pp. 3748–3754.
- J. Sola, A. Monin, M. Devy, and T. Vidal-Calleja, “Fusing monocular information in multicamera SLAM,” IEEE transactions on robotics, vol. 24, no. 5, pp. 958–968, 2008.
- M. Kaess and F. Dellaert, “Probabilistic structure matching for visual SLAM with a multi-camera rig,” Computer Vision and Image Understanding, vol. 114, no. 2, pp. 286–296, 2010.
- J. Kuo, M. Muglikar, Z. Zhang, and D. Scaramuzza, “Redesigning slam for arbitrary multi-camera systems,” in 2020 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2020.
- M. Helmberger, K. Morin, N. Kumar, D. Wang, Y. Yue, G. Cioffi, and D. Scaramuzza, “The hilti slam challenge dataset,” arXiv preprint arXiv:2109.11316, 2021.
- L. Heng, B. Choi, Z. Cui, M. Geppert, S. Hu, B. Kuan, P. Liu, R. Nguyen, Y. C. Yeo, A. Geiger, G. H. Lee, M. Pollefeys, and T. Sattler, “Project autovision: Localization and 3D scene perception for an autonomous vehicle with a multi-camera system,” arXiv, 2018.
- S. Urban and S. Hinz, “Multicol-slam-a modular real-time multi-camera slam system,” arXiv preprint arXiv:1610.07336, 2016.
- Z. Zhang and K. Zou, “MMO-SLAM: A Versatile and Accurate Multi Monocular SLAM System,” Journal of Intelligent & Robotic Systems, vol. 105, no. 3, pp. 1–23, 2022.
- G. Carrera, A. Angeli, and A. J. Davison, “Lightweight SLAM and Navigation with a Multi-Camera Rig,” European Conference on Mobile Robots, pp. 77–82, 2011.
- M. J. Tribou, A. Harmat, D. W. Wang, I. Sharf, and S. L. Waslander, “Multi-camera parallel tracking and mapping with non-overlapping fields of view,” International Journal of Robotics Research, 2015.
- L. Heng, G. H. Lee, and M. Pollefeys, “Self-calibration and visual slam with a multi-camera system on a micro aerial vehicle,” Autonomous robots, vol. 39, no. 3, pp. 259–277, 2015.
- H. Seok and J. Lim, “Rovo: Robust omnidirectional visual odometry for wide-baseline wide-fov camera systems,” in 2019 International Conference on Robotics and Automation (ICRA). IEEE, 2019.
- R. Pless, “Using many cameras as one,” in 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings., vol. 2. IEEE, 2003, pp. II–587.
- Z. Zhang, H. Rebecq, C. Forster, and D. Scaramuzza, “Benefit of large field-of-view cameras for visual odometry,” in 2016 IEEE International Conference on Robotics and Automation (ICRA).
- K. Chappellet, G. Caron, F. Kanehiro, K. Sakurada, and A. Kheddar, “Benchmarking cameras for open vslam indoors,” in 2020 25th International Conference on Pattern Recognition (ICPR). IEEE, 2021.
- H. Matsuki, L. Von Stumberg, V. Usenko, J. Stückler, and D. Cremers, “Omnidirectional DSO: Direct sparse odometry with fisheye cameras,” IEEE Robotics and Automation Letters, vol. 3, no. 4, 2018.
- S. Houben, J. Quenzel, N. Krombach, and S. Behnke, “Efficient multi-camera visual-inertial SLAM for micro aerial vehicles,” IEEE International Conference on Intelligent Robots and Systems, vol. 2016-Novem, no. October, pp. 1616–1622, 2016.
- J. Jaekel, J. G. Mangelson, S. Scherer, and M. Kaess, “A robust multi-stereo visual-inertial odometry pipeline,” IEEE International Conference on Intelligent Robots and Systems, pp. 4623–4630, 2020.
- P. Liu, M. Geppert, L. Heng, T. Sattler, A. Geiger, and M. Pollefeys, “Towards Robust Visual Odometry with a Multi-Camera System,” IEEE International Conference on Intelligent Robots and Systems, 2018.
- P. Kaveti, J. S. Nir, and H. Singh, “Towards robust vslam in dynamic environments: A light field approach,” in 2021 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI). IEEE, 2021.
- A. Li, D. Zou, and W. Yu, “Robust initialization of multi-camera slam with limited view overlaps and inaccurate extrinsic calibration,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2021, pp. 3361–3367.
- L. Zhang, D. Wisth, M. Camurri, and M. Fallon, “Balancing the budget: Feature selection and tracking for multi-camera visual-inertial odometry,” IEEE Robotics and Automation Letters, vol. 7, no. 2, 2021.
- H. Li, R. Hartley, and J.-h. Kim, “A linear approach to motion estimation using generalized camera models,” in 2008 IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 2008, pp. 1–8.
- L. Kneip, H. Li, and Y. Seo, “Upnp: An optimal o (n) solution to the absolute pose problem with universal applicability,” in European conference on computer vision. Springer, 2014, pp. 127–142.
- F. Dellaert, M. Kaess, et al., “Factor graphs for robot perception,” Foundations and Trends® in Robotics, vol. 6, no. 1-2, 2017.
- P. Furgale, J. Rehder, and R. Siegwart, “Unified temporal and spatial calibration for multi-sensor systems,” in 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2013.