Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

KDD-LOAM: Jointly Learned Keypoint Detector and Descriptors Assisted LiDAR Odometry and Mapping (2309.15394v1)

Published 27 Sep 2023 in cs.CV and cs.RO

Abstract: Sparse keypoint matching based on distinct 3D feature representations can improve the efficiency and robustness of point cloud registration. Existing learning-based 3D descriptors and keypoint detectors are either independent or loosely coupled, so they cannot fully adapt to each other. In this work, we propose a tightly coupled keypoint detector and descriptor (TCKDD) based on a multi-task fully convolutional network with a probabilistic detection loss. In particular, this self-supervised detection loss fully adapts the keypoint detector to any jointly learned descriptors and benefits the self-supervised learning of descriptors. Extensive experiments on both indoor and outdoor datasets show that our TCKDD achieves state-of-the-art performance in point cloud registration. Furthermore, we design a keypoint detector and descriptors-assisted LiDAR odometry and mapping framework (KDD-LOAM), whose real-time odometry relies on keypoint descriptor matching-based RANSAC. The sparse keypoints are further used for efficient scan-to-map registration and mapping. Experiments on KITTI dataset demonstrate that KDD-LOAM significantly surpasses LOAM and shows competitive performance in odometry.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. J. Behley and C. Stachniss, “Efficient surfel-based slam using 3d laser range data in urban environments.,” in Robotics: Science and Systems, vol. 2018, p. 59, 2018.
  2. A. Zeng, S. Song, M. Nießner, M. Fisher, J. Xiao, and T. Funkhouser, “3dmatch: Learning local geometric descriptors from rgb-d reconstructions,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1802–1811, 2017.
  3. P. Besl and N. D. McKay, “A method for registration of 3-d shapes,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 14, no. 2, pp. 239–256, 1992.
  4. H. Chen and B. Bhanu, “3d free-form object recognition in range images using local surface patches,” Pattern Recognition Letters, vol. 28, no. 10, pp. 1252–1262, 2007.
  5. Y. Zhong, “Intrinsic shape signatures: A shape descriptor for 3d object recognition,” in IEEE International conference on Computer Vision Workshops, ICCV workshops, pp. 689–696, IEEE, 2009.
  6. R. B. Rusu, N. Blodow, Z. C. Marton, and M. Beetz, “Aligning point cloud views using persistent feature histograms,” in 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 3384–3391, IEEE, 2008.
  7. R. B. Rusu, N. Blodow, and M. Beetz, “Fast point feature histograms (fpfh) for 3d registration,” in 2009 IEEE International Conference on Robotics and Automation, pp. 3212–3217, IEEE, 2009.
  8. S. Salti, F. Tombari, and L. Di Stefano, “Shot: Unique signatures of histograms for surface and texture description,” Computer Vision and Image Understanding, vol. 125, pp. 251–264, 2014.
  9. X. Bai, Z. Luo, L. Zhou, H. Fu, L. Quan, and C.-L. Tai, “D3feat: Joint learning of dense detection and description of 3d local features,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6359–6367, 2020.
  10. H. Thomas, C. R. Qi, J.-E. Deschaud, B. Marcotegui, F. Goulette, and L. J. Guibas, “Kpconv: Flexible and deformable convolution for point clouds,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 6411–6420, 2019.
  11. A. Geiger, P. Lenz, and R. Urtasun, “Are we ready for autonomous driving? the kitti vision benchmark suite,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 3354–3361, IEEE, 2012.
  12. J. Zhang and S. Singh, “Loam: Lidar odometry and mapping in real-time.,” in Robotics: Science and systems, 2014.
  13. Z. Gojcic, C. Zhou, J. D. Wegner, and A. Wieser, “The perfect match: 3d point cloud matching with smoothed densities,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5545–5554, 2019.
  14. H. Deng, T. Birdal, and S. Ilic, “Ppfnet: Global context aware local features for robust 3d point matching,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 195–205, 2018.
  15. C. R. Qi, H. Su, K. Mo, and L. J. Guibas, “Pointnet: Deep learning on point sets for 3d classification and segmentation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 652–660, 2017.
  16. C. Choy, J. Park, and V. Koltun, “Fully convolutional geometric features,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 8958–8966, 2019.
  17. S. Ao, Q. Hu, B. Yang, A. Markham, and Y. Guo, “Spinnet: Learning a general surface descriptor for 3d point cloud registration,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11753–11762, 2021.
  18. S. Huang, Z. Gojcic, M. Usvyatsov, A. Wieser, and K. Schindler, “Predator: Registration of 3d point clouds with low overlap,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4267–4276, 2021.
  19. H. Yu, F. Li, M. Saleh, B. Busam, and S. Ilic, “Cofinet: Reliable coarse-to-fine correspondences for robust point cloud registration,” Advances in Neural Information Processing Systems, vol. 34, pp. 23872–23884, 2021.
  20. Z. Qin, H. Yu, C. Wang, Y. Guo, Y. Peng, S. Ilic, D. Hu, and K. Xu, “Geotransformer: Fast and robust point cloud registration with geometric transformer,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  21. J. Li and G. H. Lee, “Usip: Unsupervised stable interest point detection from 3d point clouds,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 361–370, 2019.
  22. Z. J. Yew and G. H. Lee, “3dfeat-net: Weakly supervised local 3d features for point cloud registration,” in Proceedings of the European Conference on Computer Vision (ECCV), pp. 607–623, 2018.
  23. T. Shan and B. Englot, “Lego-loam: Lightweight and ground-optimized lidar odometry and mapping on variable terrain,” in 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4758–4765, IEEE, 2018.
  24. H. Wang, C. Wang, C.-L. Chen, and L. Xie, “F-loam: Fast lidar odometry and mapping,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4390–4396, IEEE, 2021.
  25. Y. Wu and K. He, “Group normalization,” in Proceedings of the European Conference on Computer Vision (ECCV), pp. 3–19, 2018.
  26. I. Vizzo, T. Guadagnino, B. Mersch, L. Wiesmann, J. Behley, and C. Stachniss, “Kiss-icp: In defense of point-to-point icp–simple, accurate, and robust registration if done the right way,” IEEE Robotics and Automation Letters, vol. 8, no. 2, pp. 1029–1036, 2023.
  27. P. Dellenbach, J.-E. Deschaud, B. Jacquet, and F. Goulette, “Ct-icp: Real-time elastic lidar odometry with loop closure,” in IEEE International Conference on Robotics and Automation (ICRA), pp. 5580–5586, IEEE, 2022.
  28. H. Wang, Y. Liu, Z. Dong, and W. Wang, “You only hypothesize once: Point cloud registration with rotation-equivariant descriptors,” in Proceedings of the 30th ACM International Conference on Multimedia, pp. 1630–1641, 2022.
  29. A. E. Johnson and M. Hebert, “Using spin images for efficient object recognition in cluttered 3d scenes,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 21, no. 5, pp. 433–449, 1999.
  30. M. Khoury, Q.-Y. Zhou, and V. Koltun, “Learning compact geometric features,” in Proceedings of the IEEE International Conference on Computer Vision, pp. 153–161, 2017.
  31. H. Deng, T. Birdal, and S. Ilic, “Ppf-foldnet: Unsupervised learning of rotation invariant 3d local descriptors,” in Proceedings of the European Conference on Computer Vision (ECCV), pp. 602–618, 2018.
  32. X. Chen, A. Milioto, E. Palazzolo, P. Giguere, J. Behley, and C. Stachniss, “Suma++: Efficient lidar-based semantic slam,” in 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4530–4537, IEEE, 2019.
Citations (2)

Summary

  • The paper presents a novel tightly-coupled keypoint detector and descriptor using a self-supervised probabilistic loss to improve LiDAR odometry and mapping.
  • It integrates RANSAC-based matching with sparse keypoints for efficient scan-to-map registration and enhanced real-time performance.
  • Extensive evaluations on 3DMatch and KITTI datasets demonstrate superior feature matching recall and registration accuracy compared to traditional methods.

KDD-LOAM: Jointly Learned Keypoint Detector and Descriptors Assisted LiDAR Odometry and Mapping

In the domain of robotics and 3D vision, point cloud registration plays a pivotal role, especially in applications such as SLAM and 3D reconstruction. The paper "KDD-LOAM: Jointly Learned Keypoint Detector and Descriptors Assisted LiDAR Odometry and Mapping," presents an innovative approach to improving the efficiency and robustness of point cloud registration. The authors introduce a tightly coupled keypoint detector and descriptor (TCKDD) utilizing a multi-task fully convolutional network, which integrates a probabilistic detection loss for self-supervised learning.

Key Contributions

The presented research introduces a novel approach to enhance LiDAR odometry by integrating a tightly coupled keypoint detector and descriptor. The primary contributions of the paper are:

  1. Tight Coupling of Detection and Description: The paper proposes a method to tightly couple 3D keypoint detection and description. By utilizing a self-supervised detection loss, the proposed model adapts the keypoint detector to the descriptors it learns, enhancing the self-supervised descriptor learning process.
  2. Efficient LiDAR Odometry and Mapping Framework (KDD-LOAM): The authors introduce KDD-LOAM, which leverages the learned keypoint descriptors for real-time odometry. The framework incorporates RANSAC-based keypoint descriptor matching for odometry and employs sparse keypoints for efficient scan-to-map registration and mapping.
  3. Probabilistic Detection Loss: The research formulates a novel probabilistic detection loss designed through maximum likelihood estimation, enabling the network to predict and utilize point-wise matchability for keypoint selection, concentrating learning efforts on salient regions.

Experimental Evaluation

The TCKDD framework is comprehensively evaluated on indoor (3DMatch) and outdoor (KITTI) datasets. The extensive experiments validate the state-of-the-art performance of TCKDD in point cloud registration tasks. Specifically, within indoor scenes, TCKDD demonstrates superior feature matching recall (FMR) and registration recall (RR) compared to both handcrafted and learning-based descriptors.

In outdoor scenarios on the KITTI dataset, TCKDD again shows remarkable results, achieving competitive relative translation error (RTE) and relative rotation error (RRE) performances, as well as a perfect registration recall, thereby substantiating its robustness and adaptability across varied conditions.

The KDD-LOAM system is further compared to existing LiDAR odometry and SLAM systems, such as LOAM, F-LOAM, SuMa, SuMa++, and KISS-ICP. The proposed system significantly surpasses classical approaches in terms of precision, maintaining competitive performance while also enhancing computational efficiency, especially in mapping and memory usage.

Implications and Future Work

This work presents significant implications for the robustness and efficiency of point cloud-based mapping and localization systems in both static and dynamic environments. By providing a method that tightly integrates keypoint detection with descriptor learning, the proposed approach mitigates limitations associated with separate learning strategies or handcrafted features and enhances both recall and accuracy in 3D data registration tasks.

Future developments may explore incorporating KDD-LOAM within broader SLAM frameworks, potentially enabling end-to-end learning systems that integrate loop closure detection and refined global pose graph optimization. Additionally, the framework could be extended or adapted for use with alternative sensors or within multi-sensory systems, expanding its applicability in mobile robotics and autonomous navigation.

In conclusion, the KDD-LOAM paper introduces a comprehensive and robust approach to LiDAR odometry, setting a noteworthy direction for future developments in the field of 3D computer vision and robotic perception systems.

X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com