Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multi-camera calibration with pattern rigs, including for non-overlapping cameras: CALICO (1903.06811v3)

Published 15 Mar 2019 in cs.CV

Abstract: This paper describes CALICO, a method for multi-camera calibration suitable for challenging contexts: stationary and mobile multi-camera systems, cameras without overlapping fields of view, and non-synchronized cameras. Recent approaches are roughly divided into infrastructure- and pattern-based. Infrastructure-based approaches use the scene's features to calibrate, while pattern-based approaches use calibration patterns. Infrastructure-based approaches are not suitable for stationary camera systems, and pattern-based approaches may constrain camera placement because shared fields of view or extremely large patterns are required. CALICO is a pattern-based approach, where the multi-calibration problem is formulated using rigidity constraints between patterns and cameras. We use a {\it pattern rig}: several patterns rigidly attached to each other or some structure. We express the calibration problem as that of algebraic and reprojection error minimization problems. Simulated and real experiments demonstrate the method in a variety of settings. CALICO compared favorably to Kalibr. Mean reconstruction accuracy error was $\le 0.71$ mm for real camera rigs, and $\le 1.11$ for simulated camera rigs. Code and data releases are available at \cite{tabb_amy_2019_3520866} and \url{https://github.com/amy-tabb/calico}.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. A. Tabb and M. J. Feldmann, “Data and code from: Multi-camera calibration with pattern rigs, including for non-overlapping cameras: Calico,” Dec. 2023. [Online]. Available: https://zenodo.org/doi/10.5281/zenodo.3520865
  2. Z. Zhang, “A flexible new technique for camera calibration,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 22, no. 11, pp. 1330–1334, Nov. 2000. [Online]. Available: https://doi.org/10.1109/34.888718
  3. B. A. Cook, “ACRONYM: Acronym CReatiON for You and Me,” arXiv:1903.12180 [astro-ph], Mar. 2019, arXiv: 1903.12180. [Online]. Available: http://arxiv.org/abs/1903.12180
  4. S. Garrido-Jurado, R. Muñoz-Salinas, F. Madrid-Cuevas, and M. Marín-Jiménez, “Automatic generation and detection of highly reliable fiducial markers under occlusion,” Pattern Recognition, vol. 47, no. 6, pp. 2280–2292, Jun. 2014. [Online]. Available: https://doi.org/10.1016/j.patcog.2014.01.005
  5. A. Tabb, H. Medeiros, M. J. Feldmann, and T. T. Santos, “Calibration of asynchronous camera networks: Calico,” 2019.
  6. M. Shah, R. D. Eastman, and T. Hong, “An overview of robot-sensor calibration methods for evaluation of perception systems,” in Proceedings of the Workshop on Performance Metrics for Intelligent Systems - PerMIS ’12.   College Park, Maryland: ACM Press, 2012, p. 15. [Online]. Available: http://dl.acm.org/citation.cfm?doid=2393091.2393095
  7. Li, Bo, L. Heng, K. Koser, and M. Pollefeys, “A multiple-camera system calibration toolbox using a feature descriptor-based calibration pattern.”   IEEE, Nov. 2013, pp. 1301–1307. [Online]. Available: http://ieeexplore.ieee.org/document/6696517/
  8. A. Liu, S. Marschner, and N. Snavely, “Caliber: Camera Localization and Calibration Using Rigidity Constraints,” International Journal of Computer Vision, vol. 118, no. 1, pp. 1–21, May 2016. [Online]. Available: https://doi.org/10.1007/s11263-015-0866-1
  9. P. Furgale, J. Rehder, and R. Siegwart, “Unified temporal and spatial calibration for multi-sensor systems,” in 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2013, pp. 1280–1286.
  10. J. Maye, P. Furgale, and R. Siegwart, “Self-supervised calibration for robotic systems,” in 2013 IEEE Intelligent Vehicles Symposium (IV), Jun. 2013, pp. 473–480, iSSN: 1931-0587.
  11. J. Maye, “Online Self-Calibration for Robotic Systems,” Doctoral Thesis, ETH Zurich, 2014, accepted: 2017-06-13T23:36:32Z. [Online]. Available: https://www.research-collection.ethz.ch/handle/20.500.11850/154673
  12. E. Olson, “AprilTag: A robust and flexible visual fiducial system,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA).   IEEE, May 2011, pp. 3400–3407.
  13. J. L. Schönberger and J.-M. Frahm, “Structure-from-Motion Revisited,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).   Las Vegas, NV, USA: IEEE, Jun. 2016, pp. 4104–4113. [Online]. Available: https://doi.org/10.1109/CVPR.2016.445
  14. S. Esquivel, F. Woelk, and R. Koch, “Calibration of a Multi-camera Rig from Non-overlapping Views,” in Pattern Recognition, F. A. Hamprecht, C. Schnörr, and B. Jähne, Eds.   Berlin, Heidelberg: Springer Berlin Heidelberg, 2007, vol. 4713, pp. 82–91, series Title: Lecture Notes in Computer Science. [Online]. Available: http://link.springer.com/10.1007/978-3-540-74936-3_9
  15. P. Lébraly, E. Royer, O. Ait-Aider, and M. Dhome, “Calibration of Non-Overlapping Cameras—Application to Vision-Based Robotics,” in Procedings of the British Machine Vision Conference 2010.   Aberystwyth: British Machine Vision Association, 2010, pp. 10.1–10.12. [Online]. Available: http://www.bmva.org/bmvc/2010/conference/paper10/index.html
  16. G. Carrera, A. Angeli, and A. J. Davison, “SLAM-based automatic extrinsic calibration of a multi-camera rig,” in 2011 IEEE International Conference on Robotics and Automation.   Shanghai, China: IEEE, May 2011, pp. 2652–2659. [Online]. Available: http://ieeexplore.ieee.org/document/5980294/
  17. Y. Lin, V. Larsson, M. Geppert, Z. Kukelova, M. Pollefeys, and T. Sattler, “Infrastructure-Based Multi-camera Calibration Using Radial Projections,” in Computer Vision – ECCV 2020, ser. Lecture Notes in Computer Science, A. Vedaldi, H. Bischof, T. Brox, and J.-M. Frahm, Eds.   Cham: Springer International Publishing, 2020, pp. 327–344.
  18. S. Thirthala and M. Pollefeys, “Radial Multi-focal Tensors,” International Journal of Computer Vision, vol. 96, no. 2, pp. 195–211, Jan. 2012. [Online]. Available: https://doi.org/10.1007/s11263-011-0463-x
  19. S. Mishra, S. Nagesh, S. Manglani, G. Mills, P. Chakravarty, and G. Pandey, “Look Both Ways: Bidirectional Visual Sensing for Automatic Multi-Camera Registration,” Aug. 2022, arXiv:2208.07362 [cs]. [Online]. Available: http://arxiv.org/abs/2208.07362
  20. H. Chen and S. Schwertfeger, “Heterogeneous Multi-sensor Calibration based on Graph Optimization,” in 2019 IEEE International Conference on Real-time Computing and Robotics (RCAR), Aug. 2019, pp. 158–163.
  21. T. Pollok and E. Monari, “A visual SLAM-based approach for calibration of distributed camera networks,” in 2016 13th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Aug. 2016, pp. 429–437.
  22. A. Robinson, M. Persson, and M. Felsberg, “Robust Accurate Extrinsic Calibration of Static Non-overlapping Cameras,” in Computer Analysis of Images and Patterns, ser. Lecture Notes in Computer Science, M. Felsberg, A. Heyden, and N. Krüger, Eds.   Cham: Springer International Publishing, 2017, pp. 342–353.
  23. A. Tabb and K. M. Ahmad Yousef, “Parameterizations for reducing camera reprojection error for robot-world hand-eye calibration,” in 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Sep. 2015, pp. 3030–3037. [Online]. Available: http://doi.org/10.1109/IROS.2015.7353795
  24. A. Tabb and K. M. Ahmad Yousef, “Solving the robot-world hand-eye(s) calibration problem with iterative methods,” Machine Vision and Applications, vol. 28, no. 5, pp. 569–590, Aug. 2017. [Online]. Available: https://doi.org/10.1007/s00138-017-0841-7
  25. B. Triggs, P. F. McLauchlan, R. I. Hartley, and A. W. Fitzgibbon, “Bundle Adjustment — A Modern Synthesis,” in Vision Algorithms: Theory and Practice, G. Goos, J. Hartmanis, J. van Leeuwen, B. Triggs, A. Zisserman, and R. Szeliski, Eds.   Berlin, Heidelberg: Springer Berlin Heidelberg, 2000, vol. 1883, pp. 298–372, series Title: Lecture Notes in Computer Science. [Online]. Available: http://link.springer.com/10.1007/3-540-44480-7_21
  26. G. Bradski, “The OpenCV Library,” Dr. Dobb’s Journal of Software Tools, 2000.
  27. M. Shah, “Comparing two sets of corresponding six degree of freedom data,” Computer Vision and Image Understanding, vol. 115, no. 10, pp. 1355–1362, Oct. 2011. [Online]. Available: https://doi.org/10.1016/j.cviu.2011.05.007
  28. ——, “Solving the Robot-World/Hand-Eye Calibration Problem Using the Kronecker Product,” Journal of Mechanisms and Robotics, vol. 5, no. 3, p. 031007, Jun. 2013. [Online]. Available: https://doi.org/10.1115/1.4024473
  29. S. Agarwal, K. Mierle, and Others, “Ceres solver,” http://ceres-solver.org.
  30. R. I. Hartley and P. Sturm, “Triangulation,” Computer Vision and Image Understanding, vol. 68, no. 2, pp. 146–157, Nov. 1997. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S1077314297905476
  31. R. Hartley, “In defense of the eight-point algorithm,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 19, no. 6, pp. 580–593, Jun. 1997. [Online]. Available: http://ieeexplore.ieee.org/document/601246/
  32. J. Lee, D. Hanley, and T. Bretl, “Extrinsic calibration of multiple inertial sensors from arbitrary trajectories,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 2055–2062, 2022.
Citations (2)

Summary

  • The paper introduces CALICO, a robust method using pattern rigs to calibrate multi-camera systems, even for non-overlapping and asynchronous cameras.
  • The methodology leverages rigidity constraints and minimization techniques to compute calibration parameters through iterative initialization and refinement.
  • Experimental results demonstrate that CALICO achieves efficient calibration with lower computational overhead and superior performance compared to state-of-the-art methods.

Overview of CALICO: Multi-camera Calibration Method

The research paper introduces CALICO, a robust method for the calibration of multi-camera systems, tailored to various configurations including stationary, mobile, and non-overlapping fields of view and non-synchronized camera setups. This calibration method is particularly useful in contexts such as human activity detection and complex reconstruction tasks where camera fields may not overlap, and synchronization is not always feasible.

Methodology

CALICO is a pattern-based multi-camera calibration approach that addresses the limitations of both infrastructure and existing pattern-based methodologies. Infrastructure-based solutions rely on utilizing features in real-world scenes, which are unsuitable for stationary cameras, while traditional pattern-based methods necessitate overlapping fields of view or large calibration patterns that restrict camera placement. CALICO innovatively employs a pattern rig with multiple patterns rigidly connected, allowing the calibration of diverse camera arrangements through rigidity constraints and algebraic error minimization combined with reprojection error minimization.

The methodology involves several key steps:

  1. Individual Camera Calibration: Utilizing Zhang's algorithm to acquire internal parameters by waving a calibration target and estimating extrinsic parameters through PnP pose computation for each camera.
  2. Constraint Set Formulation: The multi-camera calibration problem is cast as a set of rigidity constraints between cameras, patterns, and time labels to solve algebraic and reprojection error minimization problems.
  3. Initialization and Refinement: CALICO iterates through the constraints to apply closed-form solutions for initializing unknown variables when only one or two remain unassigned per constraint. Subsequently, the algorithm refines these by minimizing reprojection error.
  4. Data Handling and Testing: The overall method is supported by simulated datasets and real experiments, where cases ranged from intricate camera network setups (stationary or mobile) to straightforward environments like stereo rigs.

Experimental Evaluation

The paper evaluates CALICO on several datasets including simulations and real-world camera systems (both stationary and mobile) with charuco and April tag patterns. The evaluation metrics include algebraic error, reprojection root mean squared error, and reconstruction accuracy error, providing a comprehensive assessment of CALICO’s performance. The results demonstrate effective calibration with reproducible accuracy across various configurations. Crucially, CALICO's run time is efficient; for large datasets, multi-camera calibration takes merely seconds post individual camera calibration and data loading.

Comparative Analysis

CALICO's proficiency is benchmarked against the Kalibr toolbox, a recognized state-of-the-art method. Tests indicate that CALICO often yields equivalent or superior performance, especially in scenarios where Kalibr faces challenges, such as when the dataset configurations involve non-overlapping views or asynchronous captures. Simultaneously, CALICO provides robust solutions with minimal computational overhead.

Implications and Future Directions

The implications of CALICO are multifaceted, extending reliability and flexibility in multi-camera calibration tasks across disciplines, fostering advancements in areas requiring precise reconstruction from multiple viewpoints such as autonomous systems, motion capture, and surveillance. The method solidifies its niche where conventional approaches encounter limitations, promising enhanced robustness and handling diversity in camera arrangements. Future research could delve into scaling CALICO for higher-dimensional datasets, harnessing advanced optimization strategies, and exploring applications in real-time systems without pre-configured patterns.

This paper lays foundational work for pushing the boundaries of multi-camera calibration, providing both theoretical and practical strides beneficial to the scientific community and applicable to contemporary technological challenges in automated and intelligent systems.

Github Logo Streamline Icon: https://streamlinehq.com
X Twitter Logo Streamline Icon: https://streamlinehq.com