Tight Fusion of Events and Inertial Measurements for Direct Velocity Estimation (2401.09296v1)
Abstract: Traditional visual-inertial state estimation targets absolute camera poses and spatial landmark locations while first-order kinematics are typically resolved as an implicitly estimated sub-state. However, this poses a risk in velocity-based control scenarios, as the quality of the estimation of kinematics depends on the stability of absolute camera and landmark coordinates estimation. To address this issue, we propose a novel solution to tight visual-inertial fusion directly at the level of first-order kinematics by employing a dynamic vision sensor instead of a normal camera. More specifically, we leverage trifocal tensor geometry to establish an incidence relation that directly depends on events and camera velocity, and demonstrate how velocity estimates in highly dynamic situations can be obtained over short time intervals. Noise and outliers are dealt with using a nested two-layer RANSAC scheme. Additionally, smooth velocity signals are obtained from a tight fusion with pre-integrated inertial signals using a sliding window optimizer. Experiments on both simulated and real data demonstrate that the proposed tight event-inertial fusion leads to continuous and reliable velocity estimation in highly dynamic scenarios independently of absolute coordinates. Furthermore, in extreme cases, it achieves more stable and more accurate estimation of kinematics than traditional, point-position-based visual-inertial odometry.
- R. Mur-Artal, J. M. M. Montiel, and J. D. Tardos, “Orb-slam: a versatile and accurate monocular slam system,” IEEE transactions on robotics, vol. 31, no. 5, pp. 1147–1163, 2015.
- T. Qin, P. Li, and S. Shen, “Vins-mono: A robust and versatile monocular visual-inertial state estimator,” IEEE Transactions on Robotics, vol. 34, no. 4, pp. 1004–1020, 2018.
- S. Leutenegger, S. Lynen, M. Bosse, R. Siegwart, and P. Furgale, “Keyframe-based visual–inertial odometry using nonlinear optimization,” International Journal of Robotics Research (IJRR), vol. 34, no. 3, pp. 314–334, 2015.
- Y. He, J. Zhao, Y. Guo, W. He, and K. Yuan, “Pl-vio: Tightly-coupled monocular visual–inertial odometry using point and line features,” Sensors, vol. 18, no. 4, p. 1159, 2018.
- A. Pumarola, A. Vakhitov, A. Agudo, A. Sanfeliu, and F. Moreno-Noguer, “Pl-slam: Real-time monocular visual slam with points and lines,” in 2017 IEEE international conference on robotics and automation (ICRA). IEEE, 2017, pp. 4503–4508.
- Z. Min and E. Dunn, “VOLDOR+SLAM: For the times when feature-based or direct methods are not good enough,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2021.
- L. von Stumberg and D. Cremers, “DM-VIO: Delayed marginalization visual-inertial odometry,” vol. 7, no. 2, 2022, pp. 1408–1415.
- S. Weiss, R. Brockers, and L. Matthies, “4dof drift free navigation using inertial cues and optical flow,” in 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2013, pp. 4180–4186.
- Y. Zhou, G. Gallego, and S. Shen, “Event-based stereo visual odometry,” IEEE Transactions on Robotics, vol. 37, no. 5, pp. 1433–1450, 2021.
- E. Mueggler, G. Gallego, H. Rebecq, and D. Scaramuzza, “Continuous-time visual-inertial odometry for event cameras,” IEEE Transactions on Robotics, vol. 34, no. 6, pp. 1425–1440, 2018.
- Y.-F. Zuo, J. Yang, J. Chen, X. Wang, Y. Wang, and L. Kneip, “Devo: Depth-event camera visual odometry in challenging conditions,” arXiv preprint arXiv:2202.02556, 2022.
- X. Peng, W. Xu, J. Yang, and L. Kneip, “Continuous event-line constraint for closed-form velocity initialization,” in Proceedings of the British Machine Vision Conference (BMVC), 2021.
- C. Cadena, L. Carlone, H. Carrillo, Y. Latif, D. Scaramuzza, J. Neira, I. Reid, and J. J. Leonard, “Past, present, and future of simultaneous localization and mapping: Toward the robust-perception age,” IEEE Transactions on Robotics (T-RO), vol. 32, no. 6, 2016.
- D. Sterlow and S. Singh, “Motion estimation from image and intertial measurements,” International Journal of Robotics Research (IJRR), vol. 23, no. 12, pp. 1157–1195, 2004.
- T.-C. Dong-Si and A. I. Mourikis, “Motion tracking with fixed-lag smoothing: Algorithm and consistency analysis,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2011.
- A. I. Mourikis, S. I. Roumeliotis et al., “A multi-state constraint kalman filter for vision-aided inertial navigation.” in ICRA, vol. 2, 2007, p. 6.
- H. Strasdat, J. M. M. Montiel, and A. J. Davison, “Realtime monocular slam: Why filter?” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2010.
- T. Lupton and S. Sukkarieh, “Visual-inertial-aided navigation for high-dynamic motion in built environments without initial conditions,” IEEE Transactions on Robotics, vol. 28, no. 1, pp. 61–76, 2011.
- C. Forster, L. Carlone, F. Dellaert, and D. Scaramuzza, “On-manifold preintegration for real-time visual-inertial odometry,” IEEE Transactions on Robotics (T-RO), vol. 33, no. 1, pp. 1–21, 2017.
- C. Campos, R. Elvira, J. J. G. Rodríguez, J. M. Montiel, and J. D. Tardós, “Orb-slam3: An accurate open-source library for visual, visual–inertial, and multimap slam,” IEEE Transactions on Robotics, vol. 37, no. 6, pp. 1874–1890, 2021.
- X. Song, L. D. Seneviratne, K. Althoefer, and Z. Song, “Vision-based velocity estimation for unmanned ground vehicles,” International Journal of Information Acquisition, vol. 4, no. 04, pp. 303–315, 2007.
- D. Honegger, P. Greisen, L. Meier, P. Tanskanen, and M. Pollefeys, “Real-time velocity estimation based on optical flow and disparity matching,” in 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2012, pp. 5177–5182.
- D. Honegger, L. Meier, P. Tanskanen, and M. Pollefeys, “An open source and open hardware embedded metric optical flow cmos camera for indoor and outdoor applications,” in 2013 IEEE International Conference on Robotics and Automation. IEEE, 2013, pp. 1736–1741.
- S. Weiss, M. W. Achtelik, S. Lynen, M. Chli, and R. Siegwart, “Real-time onboard visual-inertial state estimation and self-calibration of mavs in unknown environments,” in 2012 IEEE international conference on robotics and automation. IEEE, 2012, pp. 957–964.
- H. Deng, U. Arif, Q. Fu, Z. Xi, Q. Quan, and K.-Y. Cai, “Visual–inertial estimation of velocity for multicopters based on vision motion constraint,” Robotics and Autonomous Systems, vol. 107, pp. 262–279, 2018.
- Z. Gao, B. Ramesh, W.-Y. Lin, P. Wang, X. Yan, and R. Zhai, “Efficient velocity estimation for mavs by fusing motion from two frontally parallel cameras,” Journal of Real-Time Image Processing, vol. 16, no. 6, pp. 2367–2378, 2019.
- J. Weng, T. S. Huang, and N. Ahuja, “Motion and structure from line correspondences; closed-form solution, uniqueness, and optimization,” IEEE Transactions on Pattern Analysis & Machine Intelligence, vol. 14, no. 03, pp. 318–336, 1992.
- R. I. Hartley, “Lines and points in three views and the trifocal tensor,” International Journal of Computer Vision, vol. 22, no. 2, pp. 125–140, 1997.
- R. G. Von Gioi, J. Jakubowicz, J.-M. Morel, and G. Randall, “Lsd: A fast line segment detector with a false detection control,” IEEE transactions on pattern analysis and machine intelligence, vol. 32, no. 4, pp. 722–732, 2008.
- G. Gallego, T. Delbrück, G. Orchard, C. Bartolozzi, B. Taba, A. Censi, S. Leutenegger, A. J. Davison, J. Conradt, K. Daniilidis et al., “Event-based vision: A survey,” IEEE transactions on pattern analysis and machine intelligence, vol. 44, no. 1, pp. 154–180, 2020.
- 2022. [Online]. Available: https://github.com/uzh-rpg/event-based_vision_resources
- D. Weikersdorfer, R. Hoffmann, and J. Conradt, “Simultaneous localization and mapping for event-based vision systems,” in International Conference on Computer Vision Systems. Springer, 2013, pp. 133–142.
- D. Weikersdorfer, D. B. Adrian, D. Cremers, and J. Conradt, “Event-based 3d slam with a depth-augmented dynamic vision sensor,” in 2014 IEEE international conference on robotics and automation (ICRA). IEEE, 2014, pp. 359–364.
- A. Censi and D. Scaramuzza, “Low-latency event-based visual odometry,” in 2014 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2014, pp. 703–710.
- E. Mueggler, B. Huber, and D. Scaramuzza, “Event-based, 6-dof pose tracking for high-speed maneuvers,” in 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2014, pp. 2761–2768.
- G. Gallego, J. E. Lund, E. Mueggler, H. Rebecq, T. Delbruck, and D. Scaramuzza, “Event-based, 6-dof camera tracking for high-speed applications,” arXiv preprint arXiv:1607.03468, vol. 2, 2016.
- W. O. Chamorro Hernández, J. Andrade-Cetto, and J. Solà Ortega, “High-speed event camera tracking,” in Proceedings of the The 31st British Machine Vision Virtual Conference, 2020, pp. 1–12.
- S. Bryner, G. Gallego, H. Rebecq, and D. Scaramuzza, “Event-based, direct camera tracking from a photometric 3d map using nonlinear optimization,” in 2019 International Conference on Robotics and Automation (ICRA). IEEE, 2019, pp. 325–331.
- G. Gallego, H. Rebecq, and D. Scaramuzza, “A unifying contrast maximization framework for event cameras, with applications to motion, depth, and optical flow estimation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 3867–3876.
- G. Gallego, M. Gehrig, and D. Scaramuzza, “Focus is all you need: Loss functions for event-based vision,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 12 280–12 289.
- B. Kueng, E. Mueggler, G. Gallego, and D. Scaramuzza, “Low-latency visual odometry using event-based feature tracks,” in 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2016, pp. 16–23.
- D. Liu, A. Parra, and T.-J. Chin, “Globally optimal contrast maximisation for event-based motion estimation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 6349–6358.
- X. Peng, L. Gao, Y. Wang, and L. Kneip, “Globally-optimal contrast maximisation for event cameras,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021.
- H. Kim, S. Leutenegger, and A. J. Davison, “Real-time 3d reconstruction and 6-dof tracking with an event camera,” in European conference on computer vision. Springer, 2016, pp. 349–364.
- H. Rebecq, T. Horstschäfer, G. Gallego, and D. Scaramuzza, “Evo: A geometric approach to event-based 6-dof parallel tracking and mapping in real time,” IEEE Robotics and Automation Letters, vol. 2, no. 2, pp. 593–600, 2016.
- A. Zhu, N. Atanasov, and K. Daniilidis, “Event-based visual inertial odometry,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
- H. Rebecq, T. Horstschaefer, and D. Scaramuzza, “Real-time visual-inertial odometry for event cameras using keyframe-based nonlinear optimization,” in Proceedings of the British Machine Vision Conference (BMVC), 2017.
- A. R. Vidal, H. Rebecq, T. Horstschaefer, and D. Scaramuzza, “Ultimate slam? combining events, images, and imu for robust visual slam in hdr and high-speed scenarios,” IEEE Robotics and Automation Letters, vol. 3, no. 2, pp. 994–1001, 2018.
- Y. Zhou, H. Li, and L. Kneip, “Canny-vo: Visual odometry with rgb-d cameras based on geometric 3-d–2-d edge alignment,” IEEE Transactions on Robotics, vol. 35, no. 1, pp. 184–199, 2018.
- J. Hidalgo-Carrió, G. Gallego, and D. Scaramuzza, “Event-aided direct sparse odometry,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 5781–5790.
- C. Le Gentil, F. Tschopp, I. Alzugaray, T. Vidal-Calleja, R. Siegwart, and J. Nieto, “Idol: A framework for imu-dvs odometry using lines,” in Proceedings of the IEEE/RSJ Conference on Intelligent Robots and Systems (IROS). IEEE, 2020, pp. 5863–5870.
- A. Zihao Zhu, N. Atanasov, and K. Daniilidis, “Event-based visual inertial odometry,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 5391–5399.
- A. I. Maqueda, A. Loquercio, G. Gallego, N. García, and D. Scaramuzza, “Event-based vision meets deep learning on steering prediction for self-driving cars,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 5419–5427.
- M. Gehrig, S. B. Shrestha, D. Mouritzen, and D. Scaramuzza, “Event-based angular velocity regression with spiking networks,” in 2020 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2020, pp. 4195–4202.
- C. Brändli, J. Strubel, S. Keller, D. Scaramuzza, and T. Delbruck, “Elised—an event-based line segment detector,” in 2016 Second International Conference on Event-based Control, Communication, and Signal Processing (EBCCSP). IEEE, 2016, pp. 1–7.
- A. Bartoli and P. Sturm, “Structure-from-motion using lines: Representation, triangulation, and bundle adjustment,” Computer Vision and Image Understanding (CVIU), vol. 100, no. 3, pp. 416–441, 2005.
- G. Zhang, J. H. Lee, J. Lim, and I. H. Suh, “Building a 3-d line-based map using stereo slam,” IEEE Transactions on Robotics (T-RO), vol. 31, no. 6, pp. 1364–1377, 2015.
- A. Bartoli and P. Sturm, “The 3d line motion matrix and alignment of line reconstructions,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), vol. 1. IEEE, 2001, pp. I–I.
- T. Stoffregen and L. Kleeman, “Event cameras, contrast maximization and reward functions: an analysis,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 12 300–12 308.
- M. A. Fischler and R. C. Bolles, “Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography,” Communications of the ACM, vol. 24, no. 6, pp. 381–395, 1981.
- P. J. Huber, “Robust estimation of a location parameter,” in Breakthroughs in statistics. Springer, 1992, pp. 492–518.
- S. Agarwal, K. Mierle, and T. C. S. Team, “Ceres Solver,” 3 2022. [Online]. Available: https://github.com/ceres-solver/ceres-solver
- T. Qin, P. Li, and S. Shen, “VINS-MONO: A robust and versatile monocular visual-inertial state estimator,” IEEE Transactions on Robotics (T-RO), vol. 34, no. 4, pp. 1004–1020, 2018.
- L. Kneip, D. Scaramuzza, and R. Siegwart, “A novel parametrization of the perspective-three-point problem for a direct computation of absolute camera position and orientation,” in CVPR 2011. IEEE, 2011, pp. 2969–2976.
- Z. Kukelova, M. Bujnak, and T. Pajdla, “Polynomial eigenvalue solutions to the 5-pt and 6-pt relative pose problems.” in BMVC, vol. 2, no. 5, 2008, p. 2008.
- J. Yang, H. Li, D. Campbell, and Y. Jia, “Go-icp: A globally optimal solution to 3d icp point-set registration,” IEEE transactions on pattern analysis and machine intelligence, vol. 38, no. 11, pp. 2241–2254, 2015.
- J. Delmerico, T. Cieslewski, H. Rebecq, M. Faessler, and D. Scaramuzza, “Are we ready for autonomous drone racing? the uzh-fpv drone racing dataset,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2019, pp. 6713–6719.
- E. Rosten and T. Drummond, “Machine learning for high-speed corner detection,” in Computer Vision–ECCV 2006: 9th European Conference on Computer Vision, Graz, Austria, May 7-13, 2006. Proceedings, Part I 9. Springer, 2006, pp. 430–443.
- B. D. Lucas and T. Kanade, “An iterative image registration technique with an application to stereo vision,” in IJCAI’81: 7th international joint conference on Artificial intelligence, vol. 2, 1981, pp. 674–679.