Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DK-SLAM: Monocular Visual SLAM with Deep Keypoint Learning, Tracking and Loop-Closing (2401.09160v2)

Published 17 Jan 2024 in cs.RO and cs.CV

Abstract: The performance of visual SLAM in complex, real-world scenarios is often compromised by unreliable feature extraction and matching when using handcrafted features. Although deep learning-based local features excel at capturing high-level information and perform well on matching benchmarks, they struggle with generalization in continuous motion scenes, adversely affecting loop detection accuracy. Our system employs a Model-Agnostic Meta-Learning (MAML) strategy to optimize the training of keypoint extraction networks, enhancing their adaptability to diverse environments. Additionally, we introduce a coarse-to-fine feature tracking mechanism for learned keypoints. It begins with a direct method to approximate the relative pose between consecutive frames, followed by a feature matching method for refined pose estimation. To mitigate cumulative positioning errors, DK-SLAM incorporates a novel online learning module that utilizes binary features for loop closure detection. This module dynamically identifies loop nodes within a sequence, ensuring accurate and efficient localization. Experimental evaluations on publicly available datasets demonstrate that DK-SLAM outperforms leading traditional and learning based SLAM systems, such as ORB-SLAM3 and LIFT-SLAM. These results underscore the efficacy and robustness of our DK-SLAM in varied and challenging real-world environments.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (39)
  1. J. J. Engel, V. Koltun, and D. Cremers, “Direct sparse odometry,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 40, pp. 611–625, 2016.
  2. X. Gao, R. Wang, N. Demmel, and D. Cremers, “Ldso: Direct sparse odometry with loop closure,” in 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2018, pp. 2198–2204.
  3. C. Forster, M. Pizzoli, and D. Scaramuzza, “Svo: Fast semi-direct monocular visual odometry,” in 2014 IEEE International Conference on Robotics and Automation (ICRA), 2014, pp. 15–22.
  4. R. Mur-Artal and J. D. Tardós, “Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras,” IEEE Transactions on Robotics, vol. 33, no. 5, pp. 1255–1262, 2017.
  5. C. Campos, R. Elvira, J. J. G. Rodríguez, J. M. M. Montiel, and J. D. Tardós, “Orb-slam3: An accurate open-source library for visual, visual–inertial, and multimap slam,” IEEE Transactions on Robotics, vol. 37, no. 6, pp. 1874–1890, 2021.
  6. Z. Teed and J. Deng, “Droid-slam: Deep visual slam for monocular, stereo, and rgb-d cameras,” in Neural Information Processing Systems, 2021.
  7. J. Tang, L. Ericson, J. Folkesson, and P. Jensfelt, “Gcnv2: Efficient correspondence prediction for real-time slam,” IEEE Robotics and Automation Letters, vol. 4, pp. 3505–3512, 2019.
  8. H. M. S. Bruno and E. Colombini, “Lift-slam: a deep-learning feature-based monocular visual slam method,” Neurocomputing, vol. 455, pp. 97–110, 2020.
  9. Y. S. Huaiyang Huang, Haoyang Ye and M. Liu, “Monocular visual odometry using learned repeatability and description,” in IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2020.
  10. Y. Lu and G. Lu, “Superthermal: Matching thermal as visible through thermal feature exploration,” IEEE Robotics and Automation Letters, vol. 6, no. 2, pp. 2690–2697, 2021.
  11. E. Rublee, V. Rabaud, K. Konolige, and G. Bradski, “Orb: An efficient alternative to sift or surf,” in 2011 International Conference on Computer Vision, 2011, pp. 2564–2571.
  12. J. Shi and C. Tomasi, “Good features to track,” 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp. 593–600, 1994.
  13. T. Qin, P. Li, and S. Shen, “Vins-mono: A robust and versatile monocular visual-inertial state estimator,” IEEE Transactions on Robotics, vol. 34, no. 4, pp. 1004–1020, 2018.
  14. D. DeTone, T. Malisiewicz, and A. Rabinovich, “Superpoint: Self-supervised interest point detection and description,” 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 337–33 712, 2017.
  15. M. Dusmanu, I. Rocco, T. Pajdla, M. Pollefeys, J. Sivic, A. Torii, and T. Sattler, “D2-net: A trainable cnn for joint description and detection of local features,” in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 8084–8093.
  16. A. Barroso-Laguna and K. Mikolajczyk, “Key.net: Keypoint detection by handcrafted and learned cnn filters revisited,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 1, 2023.
  17. Y. Tian, B. Fan, and F. Wu, “L2-net: Deep learning of discriminative patch descriptor in euclidean space,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
  18. J. Tang, J. Folkesson, and P. Jensfelt, “Geometric correspondence network for camera motion estimation,” IEEE Robotics and Automation Letters, vol. 3, no. 2, pp. 1010–1017, 2018.
  19. D. Li, X. Shi, Q. Long, S. Liu, W. Yang, F. Wang, Q. Wei, and F. Qiao, “Dxslam: A robust and efficient visual slam system with deep features,” 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4958–4965, 2020.
  20. P.-E. Sarlin, C. Cadena, R. Siegwart, and M. Dymczyk, “From coarse to fine: Robust hierarchical localization at large scale,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
  21. J. Jiang, X. Chen, W. Dai, Z. Gao, and Y. Zhang, “Thermal-inertial slam for the environments with challenging illumination,” IEEE Robotics and Automation Letters, vol. 7, no. 4, pp. 8767–8774, 2022.
  22. K. M. Yi, E. Trulls, V. Lepetit, and P. V. Fua, “Lift: Learned invariant feature transform,” in European Conference on Computer Vision, 2016.
  23. D. Galvez-López and J. D. Tardos, “Bags of binary words for fast place recognition in image sequences,” IEEE Transactions on Robotics, vol. 28, no. 5, pp. 1188–1197, 2012.
  24. M. Calonder, V. Lepetit, C. Strecha, and P. V. Fua, “Brief: Binary robust independent elementary features,” in European Conference on Computer Vision, 2010.
  25. E. Garcia-Fidalgo and A. Ortiz, “ibow-lcd: An appearance-based loop-closure detection approach using incremental bags of binary words,” IEEE Robotics and Automation Letters, vol. 3, no. 4, 2018.
  26. R. Arandjelović, P. Gronat, A. Torii, T. Pajdla, and J. Sivic, “Netvlad: Cnn architecture for weakly supervised place recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018.
  27. K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” CoRR, vol. abs/1409.1556, 2014.
  28. Y.-Y. Jau, R. Zhu, H. Su, and M. Chandraker, “Deep keypoint-based camera pose estimation with geometric constraints,” 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4950–4957, 2020.
  29. R. Kümmerle, G. Grisetti, H. Strasdat, K. Konolige, and W. Burgard, “G2o: A general framework for graph optimization,” in IEEE International Conference on Robotics and Automation, 2011, pp. 3607–3613.
  30. Y. Wang, B. Xu, W. Fan, and C. Xiang, “A robust and efficient loop closure detection approach for hybrid ground/aerial vehicles,” Drones, vol. 7, no. 2, 2023.
  31. J. Bian, W.-Y. Lin, Y. Matsushita, S.-K. Yeung, T.-D. Nguyen, and M.-M. Cheng, “Gms: Grid-based motion statistics for fast, ultra-robust feature correspondence,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 2828–2837.
  32. T.-Y. Lin, M. Maire, S. J. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick, “Microsoft coco: Common objects in context,” in European Conference on Computer Vision, 2014.
  33. A. Geiger, P. Lenz, C. Stiller, and R. Urtasun, “Vision meets robotics: The kitti dataset,” The International Journal of Robotics Research, vol. 32, pp. 1231 – 1237, 2013.
  34. M. Burri, J. Nikolic, P. Gohl, T. Schneider, J. Rehder, S. Omari, M. Achtelik, and R. Y. Siegwart, “The euroc micro aerial vehicle datasets,” The International Journal of Robotics Research, 2016.
  35. A. Geiger, J. Ziegler, and C. Stiller, “Stereoscan: Dense 3d reconstruction in real-time,” in IEEE Intelligent Vehicles Symposium, 2011.
  36. S. Umeyama, “Least-squares estimation of transformation parameters between two point patterns,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 13, pp. 376–380, 1991.
  37. J. Zubizarreta, I. Aguinaga, and J. M. M. Montiel, “Direct sparse mapping,” IEEE Transactions on Robotics, vol. 36, no. 4, 2020.
  38. H. Yue, J. Miao, Y. Yu, W. Chen, and C. Wen, “Robust loop closure detection based on bag of superpoints and graph verification,” in 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2019, pp. 3787–3793.
  39. G. Singh, M. Wu, S. Lam, and D. Van Minh, “Hierarchical loop closure detection for long-term visual slam with semantic-geometric descriptors,” in 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), 2021, pp. 2909–2916.

Summary

  • The paper introduces a deep keypoint integration approach using Model-Agnostic Meta-Learning to adapt visual features for robust SLAM.
  • The paper presents a two-stage tracking method that combines direct pose estimation with feature matching for precise localization.
  • The paper demonstrates improved loop closure detection and overall performance on KITTI and EuRoC datasets compared to ORB-SLAM3.

Introduction to DK-SLAM

Simultaneous Localization and Mapping (SLAM) technology is critical to the backbone of autonomous systems, including cars, drones, and robots, enabling them to navigate and understand their environments. Recent advances have seen the rise of deep learning techniques that have improved the performance of visual SLAM systems. However, despite significant progress, challenges persist, particularly with handcrafted visual features that falter in dynamically lit or textured environments. This necessitates the development of SLAM systems that can adapt and learn robust features to operate effectively across a range of real-world scenes.

Innovation in DK-SLAM

A key innovation of the proposed DK-SLAM framework is the integration of adaptive deep local features into the SLAM process. The system employs a Meta-Learning strategy, specifically Model-Agnostic Meta-Learning (MAML), to enable fast and efficient training of deep feature extractors. This training schema significantly enhances the system's capacity to adapt to varied scenarios without compromising the learned knowledge. In a two-stage tracking approach, the framework initially employs a direct method to roughly estimate the pose between frames, followed by a feature matching refinement for exact pose estimation. By addressing the issue of cumulative positioning error, DK-SLAM presents an online learning binary feature-based module that ensures accurate detection of loops within sequences, a crucial factor for any SLAM system's efficacy.

Scalability and Performance

Experimentation showcases DK-SLAM's superior performance, particularly when compared against notable frameworks like ORB-SLAM3. The experiments involved publicly available datasets like KITTI and EuRoC, demonstrating DK-SLAM's edge in handling realistic and diverse traffic scenarios in outdoor settings, and the challenges of indoor navigation encountered by micro aerial vehicles (MAVs). Notably, DK-SLAM's online binary BoW consistently displayed an impressive recall rate in loop closure tests, underscoring the system’s robustness.

Future Directions

While DK-SLAM paves the way for more robust and adaptable visual SLAM systems, the interplay between GPU-based online front-end processes and CPU-based back-end tasks does surface efficiency issues. Moving forward, research will focus on improving system efficiency, potentially exploring knowledge distillation techniques for parameter reduction and comprehensive optimizations within the SLAM framework. With continuous enhancements, DK-SLAM is set to advance the frontier of intelligent autonomous navigation in both familiar and unstructured environments.