Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

An Accurate and Real-time Relative Pose Estimation from Triple Point-line Images by Decoupling Rotation and Translation (2403.11639v1)

Published 18 Mar 2024 in cs.RO and cs.CV

Abstract: Line features are valid complements for point features in man-made environments. 3D-2D constraints provided by line features have been widely used in Visual Odometry (VO) and Structure-from-Motion (SfM) systems. However, how to accurately solve three-view relative motion only with 2D observations of points and lines in real time has not been fully explored. In this paper, we propose a novel three-view pose solver based on rotation-translation decoupled estimation. First, a high-precision rotation estimation method based on normal vector coplanarity constraints that consider the uncertainty of observations is proposed, which can be solved by Levenberg-Marquardt (LM) algorithm efficiently. Second, a robust linear translation constraint that minimizes the degree of the rotation components and feature observation components in equations is elaborately designed for estimating translations accurately. Experiments on synthetic data and real-world data show that the proposed approach improves both rotation and translation accuracy compared to the classical trifocal-tensor-based method and the state-of-the-art two-view algorithm in outdoor and indoor environments.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (36)
  1. R. J. Holt and A. N. Netravali, “Motion and structure from line correspondences: Some further results,” International Journal of Imaging Systems and Technology, vol. 5, no. 1, pp. 52–61, 1994.
  2. R. Gomez-Ojeda, F.-A. Moreno, D. Zuniga-Noël, D. Scaramuzza, and J. Gonzalez-Jimenez, “Pl-slam: A stereo slam system through the combination of points and line segments,” IEEE Transactions on Robotics, vol. 35, no. 3, pp. 734–746, 2019.
  3. H. Wei, F. Tang, Z. Xu, and Y. Wu, “Structural regularity aided visual-inertial odometry with novel coordinate alignment and line triangulation,” IEEE Robotics and Automation Letters, vol. 7, no. 4, pp. 10613–10620, 2022.
  4. S. Liu, Y. Yu, R. Pautrat, M. Pollefeys, and V. Larsson, “3d line mapping revisited,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 21445–21455, 2023.
  5. Y. He, J. Zhao, Y. Guo, W. He, and K. Yuan, “Pl-vio: Tightly-coupled monocular visual–inertial odometry using point and line features,” Sensors, vol. 18, no. 4, p. 1159, 2018.
  6. H. Wei, F. Tang, Z. Xu, C. Zhang, and Y. Wu, “A point-line vio system with novel feature hybrids and with novel line predicting-matching,” IEEE Robotics and Automation Letters, vol. 6, no. 4, pp. 8681–8688, 2021.
  7. Y. Liu, T. S. Huang, and O. D. Faugeras, “Determination of camera location from 2-d to 3-d line and point correspondences,” IEEE Transactions on pattern analysis and machine intelligence, vol. 12, no. 1, pp. 28–37, 1990.
  8. A. Vakhitov, L. Ferraz, A. Agudo, and F. Moreno-Noguer, “Uncertainty-aware camera pose estimation from points and lines,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4659–4668, 2021.
  9. R. Fabbri, T. Duff, H. Fan, M. H. Regan, D. d. C. d. Pinho, E. Tsigaridas, C. W. Wampler, J. D. Hauenstein, P. J. Giblin, B. Kimia, et al., “Trplp-trifocal relative pose from lines at points,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 12073–12083, 2020.
  10. D. Nistér, “An efficient solution to the five-point relative pose problem,” IEEE transactions on pattern analysis and machine intelligence, vol. 26, no. 6, pp. 756–770, 2004.
  11. H. Stewenius, C. Engels, and D. Nistér, “Recent developments on direct relative orientation,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 60, no. 4, pp. 284–294, 2006.
  12. L. Kneip, R. Siegwart, and M. Pollefeys, “Finding the exact rotation between two images independently of the translation,” in Computer Vision–ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, October 7-13, 2012, Proceedings, Part VI 12, pp. 696–709, Springer, 2012.
  13. L. Kneip and S. Lynen, “Direct optimization of frame-to-frame rotation,” in Proceedings of the IEEE International Conference on Computer Vision, pp. 2352–2359, 2013.
  14. D. Muhle, L. Koestler, N. Demmel, F. Bernard, and D. Cremers, “The probabilistic normal epipolar constraint for frame-to-frame rotation optimization under uncertain feature positions,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1819–1828, 2022.
  15. C. L. Lawson, “Contribution to the theory of linear least maximum approximation,” Ph. D. dissertation. Univ. Calif., 1961.
  16. C.-K. Chng, A. Parra, T.-J. Chin, and Y. Latif, “Monocular rotational odometry with incremental rotation averaging and loop closure,” in 2020 Digital Image Computing: Techniques and Applications (DICTA), pp. 1–8, IEEE, 2020.
  17. A. Concha, M. Burri, J. Briales, C. Forster, and L. Oth, “Instant visual odometry initialization for mobile ar,” IEEE Transactions on Visualization and Computer Graphics, vol. 27, no. 11, pp. 4226–4235, 2021.
  18. Y. He, B. Xu, Z. Ouyang, and H. Li, “A rotation-translation-decoupled solution for robust and efficient visual-inertial initialization,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 739–748, 2023.
  19. R. Hartley and A. Zisserman, Multiple view geometry in computer vision. Cambridge university press, 2003.
  20. B. Guan, P. Vasseur, and C. Demonceaux, “Trifocal tensor and relative pose estimation from 8 lines and known vertical direction,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 6001–6008, IEEE, 2022.
  21. P. Hruby, T. Duff, A. Leykin, and T. Pajdla, “Learning to solve hard minimal problems,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5532–5542, 2022.
  22. Y. Ding, C.-H. Chien, V. Larsson, K. Åström, and B. Kimia, “Minimal solutions to generalized three-view relative pose problem,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 8156–8164, 2023.
  23. A. Morgan, Solving polynomial systems using continuation for engineering and scientific problems. SIAM, 2009.
  24. T. Duff, K. Kohn, A. Leykin, and T. Pajdla, “Plmp-point-line minimal problems in complete multi-view visibility,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1675–1684, 2019.
  25. R. Fabbri, T. Duff, H. Fan, M. H. Regan, D. d. C. de Pinho, E. Tsigaridas, C. W. Wampler, J. D. Hauenstein, P. J. Giblin, B. Kimia, et al., “Trifocal relative pose from lines at points,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
  26. Z. Xu, H. Wei, F. Tang, Y. Zhang, Y. Wu, G. Ma, S. Wu, and X. Jin, “Plpl-vio: a novel probabilistic line measurement model for point-line-based visual-inertial odometry,” in 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5211–5218, IEEE, 2023.
  27. J. K. Uhlmann, Dynamic map building and localization: new theoretical foundations. PhD thesis, University of Oxford, 1995.
  28. Q. Cai, L. Zhang, Y. Wu, W. Yu, and D. Hu, “A pose-only solution to visual reconstruction and navigation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 1, pp. 73–86, 2021.
  29. R. I. Hartley, “In defense of the eight-point algorithm,” IEEE Transactions on pattern analysis and machine intelligence, vol. 19, no. 6, pp. 580–593, 1997.
  30. M. A. Fischler and R. C. Bolles, “Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography,” Communications of the ACM, vol. 24, no. 6, pp. 381–395, 1981.
  31. A. Geiger, P. Lenz, and R. Urtasun, “Are we ready for autonomous driving? the kitti vision benchmark suite,” in 2012 IEEE conference on computer vision and pattern recognition, pp. 3354–3361, IEEE, 2012.
  32. M. Burri, J. Nikolic, P. Gohl, T. Schneider, J. Rehder, S. Omari, M. W. Achtelik, and R. Siegwart, “The euroc micro aerial vehicle datasets,” The International Journal of Robotics Research, vol. 35, no. 10, pp. 1157–1163, 2016.
  33. E. Rosten and T. Drummond, “Machine learning for high-speed corner detection,” in Computer Vision–ECCV 2006: 9th European Conference on Computer Vision, Graz, Austria, May 7-13, 2006. Proceedings, Part I 9, pp. 430–443, Springer, 2006.
  34. B. D. Lucas and T. Kanade, “An iterative image registration technique with an application to stereo vision,” in IJCAI’81: 7th international joint conference on Artificial intelligence, vol. 2, pp. 674–679, 1981.
  35. R. G. Von Gioi, J. Jakubowicz, J.-M. Morel, and G. Randall, “Lsd: A line segment detector,” Image Processing On Line, vol. 2, pp. 35–55, 2012.
  36. L. Zhang and R. Koch, “An efficient and robust line segment matching approach based on lbd descriptor and pairwise geometric consistency,” Journal of visual communication and image representation, vol. 24, no. 7, pp. 794–805, 2013.
Citations (2)

Summary

  • The paper introduces RT²PL, a decoupled approach that separately estimates high-precision rotation and robust translation.
  • It employs normal vector coplanarity constraints for superior rotation accuracy, effectively overcoming challenges in weak-textured settings.
  • Extensive experiments on synthetic and real-world data show marked improvements over traditional and state-of-the-art pose estimation methods.

Advanced Pose Estimation from Triple Point-Line Images with Decoupled Rotation and Translation

Overview

A novel methodology for three-view relative pose estimation, leveraging both point and line features within man-made environments, is proposed. This paper illustrates an approach that significantly enhances the accuracy of pose estimation in scenarios characterized by weak-textured surfaces by utilizing the complementary information provided by line features alongside point features. Unlike existing methodologies that mix rotation and translation parameters, leading to potential inaccuracies and inefficiencies, this research introduces a decoupled estimation of rotation and translation. The proposed algorithm efficiently solves for high-precision rotation through normal vector coplanarity constraints and accurately estimates translations via a uniquely designed robust linear translation constraint. The efficacy of this method has been demonstrated through comprehensive experiments on both synthetic and real-world data, showing marked improvements over classical and state-of-the-art algorithms in various environments.

Key Contributions

  • Decoupled Estimation Approach: A significant leap from traditional methods, this research presents RT2^2PL, a method that decouples the estimation of rotation and translation, addressing the challenges posed by pure rotation and planar degeneracies effectively.
  • Normal Vector Coplanarity Constraints for Rotation: High-precision rotation estimation is achieved using coplanarity constraints. Specifically, two forms of constraints are discussed - minimal eigenvalue form and eigenvalue multiplication form, with the latter demonstrating superior convergence and resilience to initial guess inaccuracies.
  • Robust Linear Translation Constraint: An innovative translation constraint is introduced, minimizing the impact of rotation estimation errors and observation noise, thus enhancing the accuracy of translation estimates.
  • Integration of Point and Line Features: The method capitalizes on the strengths of both point and line features, significantly improving pose estimation accuracy, especially in environments with weak textures where line features prevail.

Practical and Theoretical Implications

The algorithm presents a robust framework for pose estimation that could significantly benefit Visual Odometry (VO) and Structure-from-Motion (SfM) systems, especially in man-made environments characterized by weak textures. By providing a more accurate initial pose using decoupled rotation and translation estimates, the research paves the way for more reliable and efficient VO and SfM systems that can function effectively even in challenging conditions. Furthermore, the insights gained from the decoupled estimation approach and the integration of point-line features offer promising directions for future research in enhancing pose estimation methodologies.

Speculations on Future Developments

The successful application of RT2^2PL in enhancing accuracy underlines the potential of integrating additional types of features and employing this approach in more complex scenarios, such as dynamic environments. Future studies could explore the extension of this methodology to Visual-Inertial Odometry (VIO) systems, potentially leading to significant advancement in autonomous navigation technologies, including drones and autonomous vehicles operating in urban settings. Moreover, further examination into optimizing the computational efficiency of the method could facilitate its application in real-time systems that require rapid pose estimation.

Conclusion

The RT2^2PL method introduced in this paper marks a significant advancement in the field of pose estimation, providing a novel solution that accurately and efficiently solves the three-view relative pose estimation problem by decoupling the rotation and translation estimates. The algorithm not only demonstrates superior performance in terms of accuracy and efficiency compared to existing methods but also opens promising avenues for future research and developments in advanced pose estimation techniques.

X Twitter Logo Streamline Icon: https://streamlinehq.com