360ORB-SLAM: A Visual SLAM System for Panoramic Images with Depth Completion Network (2401.10560v1)
Abstract: To enhance the performance and effect of AR/VR applications and visual assistance and inspection systems, visual simultaneous localization and mapping (vSLAM) is a fundamental task in computer vision and robotics. However, traditional vSLAM systems are limited by the camera's narrow field-of-view, resulting in challenges such as sparse feature distribution and lack of dense depth information. To overcome these limitations, this paper proposes a 360ORB-SLAM system for panoramic images that combines with a depth completion network. The system extracts feature points from the panoramic image, utilizes a panoramic triangulation module to generate sparse depth information, and employs a depth completion network to obtain a dense panoramic depth map. Experimental results on our novel panoramic dataset constructed based on Carla demonstrate that the proposed method achieves superior scale accuracy compared to existing monocular SLAM methods and effectively addresses the challenges of feature association and scale ambiguity. The integration of the depth completion network enhances system stability and mitigates the impact of dynamic elements on SLAM performance.
- A. J. Davison, I. D. Reid, N. D. Molton and O. Stasse, “MonoSLAM: Real-Time Single Camera SLAM,” in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 29, no. 6, pp. 1052-1067, 2007.
- Z. Zhang, H. Rebecq, C. Forster and D. Scaramuzza, “Benefit of large field-of-view cameras for visual odometry,” 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, 2016, pp. 801-808.
- R. Mur-Artal and J. D. Tardós, “ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras,” in IEEE Transactions on Robotics, vol. 33, no. 5, pp. 1255-1262, Oct. 2017.
- J. Engel, V. Koltun and D. Cremers, “Direct Sparse Odometry,” in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 40, no. 3, pp. 611-625, 1 March 2018.
- J. Engel, T. Sch ops, and D. Cremers, “Lsd-slam: Large-scale direct monocular slam,” in European conference on computer vision. Springer, 2014, pp. 834–849.
- S. Sumikura, M. Shibuya, and K. Sakurada, “OpenvSLAM: a versatile visual slam framework,” in Proceedings of the 27th ACM International Conference on Multimedia, 2019, pp. 2292–2295.
- H. Matsuki, L. von Stumberg, V. Usenko, J. Stückler and D. Cremers, “Omnidirectional DSO: Direct Sparse Odometry With Fisheye Cameras,” in IEEE Robotics and Automation Letters, vol. 3, no. 4, pp. 3693-3700, Oct. 2018.
- C. Campos, R. Elvira, J. J. G. Rodríguez, J. M. M. Montiel and J. D. Tardós, “ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual–Inertial, and Multimap SLAM,” in IEEE Transactions on Robotics, vol. 37, no. 6, pp. 1874-1890, Dec. 2021.
- J. Kannala and S. S. Brandt, “A generic camera model and calibration method for conventional, wide-angle, and fish-eye lenses,” in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 28, no. 8, pp. 1335-1340, Aug. 2006.
- A. Eldesokey, M. Felsberg and F. S. Khan, “Confidence Propagation through CNNs for Guided Sparse Depth Regression,” in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 42, no. 10, pp. 2423-2436, 1 Oct. 2020.