Affine Correspondences between Multi-Camera Systems for Relative Pose Estimation (2306.12996v1)
Abstract: We present a novel method to compute the relative pose of multi-camera systems using two affine correspondences (ACs). Existing solutions to the multi-camera relative pose estimation are either restricted to special cases of motion, have too high computational complexity, or require too many point correspondences (PCs). Thus, these solvers impede an efficient or accurate relative pose estimation when applying RANSAC as a robust estimator. This paper shows that the 6DOF relative pose estimation problem using ACs permits a feasible minimal solution, when exploiting the geometric constraints between ACs and multi-camera systems using a special parameterization. We present a problem formulation based on two ACs that encompass two common types of ACs across two views, i.e., inter-camera and intra-camera. Moreover, the framework for generating the minimal solvers can be extended to solve various relative pose estimation problems, e.g., 5DOF relative pose estimation with known rotation angle prior. Experiments on both virtual and real multi-camera systems prove that the proposed solvers are more efficient than the state-of-the-art algorithms, while resulting in a better relative pose accuracy. Source code is available at https://github.com/jizhaox/relpose-mcs-depth.
- D. Nistér, “An efficient solution to the five-point relative pose problem,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 26, no. 6, pp. 756–777, 2004.
- H. Stewénius, M. Oskarsson, K. Aström, and D. Nistér, “Solutions to minimal generalized relative pose problems,” in Workshop on Omnidirectional Vision in conjunction with ICCV, 2005, pp. 1–8.
- D. Scaramuzza and F. Fraundorfer, “Visual odometry: The first 30 years and fundamentals,” IEEE Robotics & Automation Magazine, vol. 18, no. 4, pp. 80–92, 2011.
- L. Kneip and H. Li, “Efficient computation of relative pose for multi-camera systems,” in IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 446–453.
- J. L. Schönberger and J.-M. Frahm, “Structure-from-motion revisited,” in IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 4104–4113.
- C. Häne, L. Heng, G. H. Lee, F. Fraundorfer, P. Furgale, T. Sattler, and M. Pollefeys, “3D visual perception for self-driving cars using a multi-camera system: Calibration, mapping, localization, and obstacle detection,” Image and Vision Computing, vol. 68, pp. 14–27, 2017.
- L. Heng, B. Choi, Z. Cui, M. Geppert, S. Hu, B. Kuan, P. Liu, R. Nguyen, Y. C. Yeo, A. Geiger, G. H. Lee, M. Pollefeys, and T. Sattler, “Project AutoVision: Localization and 3D scene perception for an autonomous vehicle with a multi-camera system,” in IEEE International Conference on Robotics and Automation, 2019, pp. 4695–4702.
- K. Alyousefi and J. Ventura, “Multi-camera motion estimation with affine correspondences,” in International Conference on Image Analysis and Recognition, 2020, pp. 417–431.
- E. Martyushev and B. Li, “Efficient relative pose estimation for cameras and generalized cameras in case of known relative rotation angle,” Journal of Mathematical Imaging and Vision, vol. 62, pp. 1076–1086, 2020.
- B. Guan, J. Zhao, D. Barath, and F. Fraundorfer, “Minimal cases for computing the generalized relative pose using affine correspondences,” in IEEE International Conference on Computer Vision, 2021, pp. 6068–6077.
- M. D. Grossberg and S. K. Nayar, “A general imaging model and a method for finding its parameters,” in IEEE International Conference on Computer Vision, vol. 2. IEEE, 2001, pp. 108–115.
- P. Sturm and S. Ramalingam, “A generic concept for camera calibration,” in European Conference on Computer Vision, vol. 3022. Springer-Verlag, 2004, pp. 1–13.
- P. Miraldo, H. Araujo, and J. Queiró, “Point-based calibration using a parametric representation of the general imaging model,” in IEEE International Conference on Computer Vision. IEEE, 2011, pp. 2304–2311.
- R. Pless, “Using many cameras as one,” in IEEE Conference on Computer Vision and Pattern Recognition, 2003, pp. 1–7.
- M. A. Fischler and R. C. Bolles, “Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography,” Communications of the ACM, vol. 24, no. 6, pp. 381–395, 1981.
- H. Li, R. Hartley, and J.-h. Kim, “A linear approach to motion estimation using generalized camera models,” in IEEE Conference on Computer Vision and Pattern Recognition, 2008, pp. 1–8.
- J.-H. Kim, H. Li, and R. Hartley, “Motion estimation for nonoverlapping multicamera rigs: Linear algebraic and L∞subscript𝐿L_{\infty}italic_L start_POSTSUBSCRIPT ∞ end_POSTSUBSCRIPT geometric solutions,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 32, no. 6, pp. 1044–1059, 2009.
- J. Lim, N. Barnes, and H. Li, “Estimating relative camera motion from the antipodal-epipolar constraint,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 32, no. 10, pp. 1907–1914, 2010.
- J. Ventura, C. Arth, and V. Lepetit, “An efficient minimal solution for multi-camera motion,” in IEEE International Conference on Computer Vision, 2015, pp. 747–755.
- L. Kneip, C. Sweeney, and R. Hartley, “The generalized relative pose and scale problem: View-graph fusion via 2D-2D registration,” in IEEE Winter Conference on Applications of Computer Vision, 2016, pp. 1–9.
- J. Zhao, W. Xu, and L. Kneip, “A certifiably globally optimal solution to generalized essential matrix estimation,” in IEEE Conference on Computer Vision and Pattern Recognition, 2020, pp. 12 034–12 043.
- G. H. Lee, F. Faundorfer, and M. Pollefeys, “Motion estimation for self-driving cars with a generalized camera,” in IEEE Conference on Computer Vision and Pattern Recognition, 2013, pp. 2746–2753.
- G. H. Lee, M. Pollefeys, and F. Fraundorfer, “Relative pose estimation for a multi-camera system with known vertical direction,” in IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 540–547.
- C. Sweeney, J. Flynn, and M. Turk, “Solving for relative pose with a partially known rotation is a quadratic eigenvalue problem,” in International Conference on 3D Vision, 2014, pp. 483–490.
- C. Sweeney, L. Kneip, T. Hollerer, and M. Turk, “Computing similarity transformations from only image correspondences,” in IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 3305–3313.
- L. Liu, H. Li, Y. Dai, and Q. Pan, “Robust and efficient relative pose with a multi-camera system for autonomous driving in highly dynamic environments,” IEEE Transactions on Intelligent Transportation Systems, vol. 19, no. 8, pp. 2432–2444, 2017.
- B. Li, L. Heng, G. H. Lee, and M. Pollefeys, “A 4-point algorithm for relative pose estimation of a calibrated camera with a known relative rotation angle,” in IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2013, pp. 1595–1601.
- B. Li, E. Martyushev, and G. H. Lee, “Relative pose estimation of calibrated cameras with known SE(3)SE3\mathrm{SE}(3)roman_SE ( 3 ) invariants,” in European Conference on Computer Vision, 2020, pp. 215–231.
- J. Bentolila and J. M. Francos, “Conic epipolar constraints from affine correspondences,” Computer Vision and Image Understanding, vol. 122, pp. 105–114, 2014.
- C. Raposo and J. P. Barreto, “Theory and practice of structure-from-motion using affine correspondences,” in IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 5470–5478.
- L. Hajder and D. Barath, “Relative planar motion for vehicle-mounted cameras from a single affine correspondence,” in IEEE International Conference on Robotics and Automation, 2020, pp. 8651–8657.
- D. Barath, M. Polic, W. FÃűrstner, T. Sattler, T. Pajdla, and Z. Kukelova, “Making affine correspondences work in camera geometry computation,” in European Conference on Computer Vision, 2020, pp. 723–740.
- I. Eichhardt and D. Barath, “Relative pose from deep learned depth and a single affine correspondence,” in European Conference on Computer Vision, 2020, pp. 627–644.
- B. Guan, J. Zhao, D. Barath, and F. Fraundorfer, “Efficient recovery of multi-camera motion from two affine correspondences,” in IEEE International Conference on Robotics and Automation, 2021, pp. 1305–1311.
- B. Guan and J. Zhao, “Affine correspondences between multi-camera systems for 6dof relative pose estimation,” in European Conference on Computer Vision, 2022, pp. 634–650.
- J.-H. Kim, R. Hartley, J.-M. Frahm, and M. Pollefeys, “Visual odometry for non-overlapping views using second-order cone programming,” in Asian Conference on Computer Vision, 2007, pp. 353–362.
- B. Clipp, J.-H. Kim, J.-M. Frahm, M. Pollefeys, and R. Hartley, “Robust 6dof motion estimation for non-overlapping, multi-camera systems,” in IEEE Workshop on Applications of Computer Vision. IEEE, 2008, pp. 1–8.
- D. Barath and L. Hajder, “Efficient recovery of essential matrix from two affine correspondences,” IEEE Transactions on Image Processing, vol. 27, no. 11, pp. 5328–5337, 2018.
- H. Stewénius, D. Nistér, F. Kahl, and F. Schaffalitzky, “A minimal solution for relative pose with unknown focal length,” in IEEE Conference on Computer Vision and Pattern Recognition, 2005, pp. 789–794.
- E. Zheng and C. Wu, “Structure from motion using structure-less resection,” in IEEE International Conference on Computer Vision, 2015, pp. 2075–2083.
- J. Zhao, L. Kneip, Y. He, and J. Ma, “Minimal case relative pose computation using ray-point-ray features,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 42, no. 5, pp. 1176–1190, 2020.
- I. Eichhardt and D. Chetverikov, “Affine correspondences between central cameras for rapid relative pose estimation,” in European Conference on Computer Vision, 2018, pp. 482–497.
- D. R. Grayson and M. E. Stillman, “Macaulay 2, a software system for research in algebraic geometry,” https://faculty.math.illinois.edu/Macaulay2/, 2002.
- V. Larsson, K. Aström, and M. Oskarsson, “Efficient solvers for minimal problems by syzygy-based reduction,” in IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 820–828.
- M. Byröd, K. Josephson, and K. Aström, “Fast and stable polynomial equation solving and its application to computer vision,” International Journal of Computer Vision, vol. 84, no. 3, pp. 237–256, 2009.
- L. Kneip and P. Furgale, “OpenGV: A unified and generalized approach to real-time calibrated geometric vision,” in IEEE International Conference on Robotics and Automation, 2014, pp. 12 034–12 043.
- B. Guan, J. Zhao, D. Barath, and F. Fraundorfer, “Minimal solvers for relative pose estimation of multi-camera systems using affine correspondences,” International Journal of Computer Vision, vol. 131, no. 1, pp. 324–345, 2023.
- D. Barath and Z. Kukelova, “Homography from two orientation-and scale-covariant features,” in IEEE International Conference on Computer Vision, 2019, pp. 1091–1099.
- A. Geiger, P. Lenz, C. Stiller, and R. Urtasun, “Vision meets robotics: The KITTI dataset,” The International Journal of Robotics Research, vol. 32, no. 11, pp. 1231–1237, 2013.
- H. Caesar, V. Bankiti, A. H. Lang, S. Vora, V. E. Liong, Q. Xu, A. Krishnan, Y. Pan, G. Baldan, and O. Beijbom, “nuScenes: A multimodal dataset for autonomous driving,” in IEEE Conference on Computer Vision and Pattern Recognition, 2020, pp. 11 621–11 631.
- M. Burri, J. Nikolic, P. Gohl, T. Schneider, J. Rehder, S. Omari, M. W. Achtelik, and R. Siegwart, “The EuRoC micro aerial vehicle datasets,” The International Journal of Robotics Research, vol. 35, no. 10, pp. 1157–1163, 2016.
- J.-M. Morel and G. Yu, “ASIFT: A new framework for fully affine invariant image comparison,” SIAM Journal on Imaging Sciences, vol. 2, no. 2, pp. 438–469, 2009.
- J. Matas, O. Chum, M. Urban, and T. Pajdla, “Robust wide-baseline stereo from maximally stable extremal regions,” Image and Vision Computing, vol. 22, no. 10, pp. 761–767, 2004.
- J. Sturm, N. Engelhard, F. Endres, W. Burgard, and D. Cremers, “A benchmark for the evaluation of RGB-D SLAM systems,” in IEEE/RSJ International Conference on Intelligent Robots and Systems, 2012, pp. 573–580.