Papers
Topics
Authors
Recent
Search
2000 character limit reached

Event-Based Visual Odometry on Non-Holonomic Ground Vehicles

Published 17 Jan 2024 in cs.CV and cs.RO | (2401.09331v1)

Abstract: Despite the promise of superior performance under challenging conditions, event-based motion estimation remains a hard problem owing to the difficulty of extracting and tracking stable features from event streams. In order to robustify the estimation, it is generally believed that fusion with other sensors is a requirement. In this work, we demonstrate reliable, purely event-based visual odometry on planar ground vehicles by employing the constrained non-holonomic motion model of Ackermann steering platforms. We extend single feature n-linearities for regular frame-based cameras to the case of quasi time-continuous event-tracks, and achieve a polynomial form via variable degree Taylor expansions. Robust averaging over multiple event tracks is simply achieved via histogram voting. As demonstrated on both simulated and real data, our algorithm achieves accurate and robust estimates of the vehicle's instantaneous rotational velocity, and thus results that are comparable to the delta rotations obtained by frame-based sensors under normal conditions. We furthermore significantly outperform the more traditional alternatives in challenging illumination scenarios. The code is available at \url{https://github.com/gowanting/NHEVO}.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (61)
  1. Asynchronous corner detection and tracking for event cameras in real time. IEEE Robotics and Automation Letters, 3(4):3177–3184, 2018.
  2. Event-based, direct camera tracking from a photometric 3d map using nonlinear optimization. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), pages 325–331. IEEE, 2019.
  3. Orb-slam3: An accurate open-source library for visual, visual-inertial and multi-map slam. arXiv preprint arXiv:2007.11898, 2020.
  4. Low-latency event-based visual odometry. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), pages 703–710. IEEE, 2014.
  5. High-speed event camera tracking. In Proceedings of the British Machine Vision Conference (BMVC), 2020.
  6. Lsd-slam: Large-scale direct monocular slam. In Proceedings of the European Conference on Computer Vision (ECCV), 2014.
  7. Direct sparse odometry. IEEE transactions on pattern analysis and machine intelligence, 40(3):611–625, 2017.
  8. SVO: Fast semi-direct monocular visual odometry. In IEEE International Conference on Robotics and Automation (ICRA), 2014.
  9. A minimal case solution to the calibrated relative pose problem for the case of two known orientation angles. In Proceedings of the European Conference on Computer Vision (ECCV), 2010.
  10. Accurate angular velocity estimation with an event camera. IEEE Robotics and Automation Letters, 2(2):632–639, 2017.
  11. Event-based, 6-dof camera tracking for high-speed applications. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 40(2):2402–2412, 2016.
  12. A unifying contrast maximization framework for event cameras, with applications to motion, depth, and optical flow estimation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3867–3876, 2018.
  13. Focus is all you need: Loss functions for event-based vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 12280–12289, 2019.
  14. Event-based vision: A survey. IEEE transactions on pattern analysis and machine intelligence, 44(1):154–180, 2020.
  15. Video to events: Recycling video datasets for event cameras. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3586–3595, 2020.
  16. Vision meets robotics: The kitti dataset. The International Journal of Robotics Research, 32(11):1231–1237, 2013.
  17. Localization and matching using the planar trifocal tensor with bearing-only data. 24(2):494–501, 2008.
  18. R.I. Hartley. Projective reconstruction from line correspondences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 1994.
  19. Multiple View Geometry in Computer Vision. Cambridge University Press, New York, NY, USA, 2004. second edition.
  20. Motion estimation for self-driving cars with a generalized camera. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2746–2753, 2013.
  21. Event-aided direct sparse odometry. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5781–5790, 2022.
  22. Motion estimation of non-holonomic ground vehicles from a single feature correspondence measured over n views. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, USA, 2019.
  23. B-splines for purely vision-based localization and mapping on non-holonomic ground vehicles. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2021.
  24. Real-time 3d reconstruction and 6-dof tracking with an event camera. In Proceedings of the European Conference on Computer Vision (ECCV), pages 349–364. Springer, 2016.
  25. Parallel tracking and mapping for small ar workspaces. In 2007 6th IEEE and ACM international symposium on mixed and augmented reality, pages 225–234. IEEE, 2007.
  26. Parallel tracking and mapping on a camera phone. In 2009 8th IEEE International Symposium on Mixed and Augmented Reality, pages 83–86. IEEE, 2009.
  27. TANDEM: Tracking and Dense Mapping in Real-time using Deep Multi-view Stereo. In In Conference on Robot Learning (CoRL), 2021.
  28. Globally optimal contrast maximisation for event-based motion estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 6349–6358, 2020.
  29. H Christopher Longuet-Higgins. A computer algorithm for reconstructing a scene from two projections. Nature, 293(5828):133–135, 1981.
  30. Voldor-slam: For the times when feature-based or direct methods are not good enough. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2021.
  31. Event-based, 6-dof pose tracking for high-speed maneuvers. In Proceedings of the IEEE/RSJ Conference on Intelligent Robots and Systems (IROS), pages 2761–2768. IEEE, 2014.
  32. Continuous-time visual-inertial odometry for event cameras. IEEE Transactions on Robotics (T-RO), 34(6):1425–1440, 2018.
  33. Montiel J. M. M. Mur-Artal, Raúl and Juan D. Tardós. ORB-SLAM: a versatile and accurate monocular SLAM system. IEEE Transactions on Robotics, 31(5):1147–1163, 2015.
  34. ORB-SLAM2: An open-source slam system for monocular, stereo, and rgb-d cameras. IEEE Transactions on Robotics, 33(5):1255–1262, 2017.
  35. Orb-slam: a versatile and accurate monocular slam system. IEEE transactions on robotics, 31(5):1147–1163, 2015.
  36. David Nistér. An efficient solution to the five-point relative pose problem. IEEE transactions on pattern analysis and machine intelligence, 26(6):756–770, 2004.
  37. Visual odometry. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004., pages I–I. Ieee, 2004.
  38. Articulated multi-perspective cameras and their application to truck motion estimation. In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 2052–2059. IEEE, 2019.
  39. Globally-optimal contrast maximisation for event cameras. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(7):3479–3495, 2021a.
  40. Continuous event-line constraint for closed-form velocity initialization. In Proceedings of the British Machine Vision Conference (BMVC), 2021b.
  41. Relative pose estimation for instrumented, calibrated imaging platforms. In DICTA, pages 601–612. Sydney, Australia, 2003.
  42. VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator. IEEE Transactions on Robotics (T-RO), 34(4):1004–1020, 2018.
  43. Evo: A geometric approach to event-based 6-dof parallel tracking and mapping in real time. IEEE Robotics and Automation Letters, 2(2):593–600, 2016.
  44. Real-time visual-inertial odometry for event cameras using keyframe-based nonlinear optimization. In Proceedings of the British Machine Vision Conference (BMVC), 2017.
  45. Davide Scaramuzza. 1-point-ransac structure from motion for vehicle-mounted cameras by exploiting non-holonomic constraints. International Journal of Computer Vision (IJCV), 95(1):74–85, 2011.
  46. Visual odometry [tutorial]. IEEE robotics & automation magazine, 18(4):80–92, 2011.
  47. Absolute scale in structure from motion from a single vehicle mounted camera by exploiting nonholonomic constraints. In 2009 IEEE 12th international conference on computer vision, pages 1413–1419. IEEE, 2009a.
  48. Real-time monocular visual odometry for on-road vehicles with 1-point ransac. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), pages 4293–4299. Ieee, 2009b.
  49. Lego-loam: Lightweight and ground-optimized lidar odometry and mapping on variable terrain. In IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 4758–4765. IEEE, 2018.
  50. Recent developments on direct relative orientation. ISPRS Journal of Photogrammetry and Remote Sensing, 60(4):284–294, 2006.
  51. Event cameras, contrast maximization and reward functions: an analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 12300–12308, 2019.
  52. DROID-SLAM: Deep Visual SLAM for Monocular, Stereo, and RGB-D Cameras. Advances in neural information processing systems, 2021.
  53. Ultimate slam? combining events, images, and imu for robust visual slam in hdr and high-speed scenarios. IEEE Robotics and Automation Letters, 3(2):994–1001, 2018.
  54. Simultaneous localization and mapping for event-based vision systems. In Computer Vision Systems: 9th International Conference, ICVS 2013, St. Petersburg, Russia, July 16-18, 2013. Proceedings 9, pages 133–142. Springer, 2013.
  55. Event-based 3d slam with a depth-augmented dynamic vision sensor. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), pages 359–364. IEEE, 2014.
  56. Tight fusion of events and inertial measurements for direct velocity estimation. IEEE Transactions on Robotics, pages 1–17, 2023.
  57. Event-based stereo visual odometry. IEEE Transactions on Robotics, 37(5):1433–1450, 2021.
  58. Neuromorphic visual odometry system for intelligent vehicle application with bio-inspired vision sensor. In 2019 IEEE International Conference on Robotics and Biomimetics (ROBIO), pages 2225–2232. IEEE, 2019.
  59. Event-based visual inertial odometry. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 5391–5399, 2017.
  60. Devo: Depth-event camera visual odometry in challenging conditions. In 2022 International Conference on Robotics and Automation (ICRA), pages 2179–2185. IEEE, 2022.
  61. Cross-modal semi-dense 6-dof tracking of an event camera in challenging conditions. IEEE Transactions on Robotics, pages 1–17, 2024.

Summary

  • The paper introduces a purely event-based visual odometry system leveraging non-holonomic Ackermann models for robust motion estimation.
  • It employs polynomial formulations and Taylor series expansions to adapt traditional n-linear constraints to asynchronous event data.
  • Histogram voting and rank minimization techniques enhance feature tracking accuracy under challenging lighting conditions.

Event-Based Visual Odometry on Non-Holonomic Ground Vehicles

The study by Xu et al. addresses the complexities inherent in achieving effective visual odometry (VO) using event-based cameras on non-holonomic ground vehicles. Traditional visual odometry techniques using regular cameras face significant performance hurdles in high-dynamic and high-dynamic-range conditions due to motion blur and overexposure. Event-based cameras, which capture changes in illumination asynchronously on a pixel-by-pixel basis, offer a promising alternative by providing high dynamic range and rapid response times. However, processing event streams for stable feature extraction and tracking poses substantial challenges.

The authors propose a purely event-based visual odometry solution by leveraging the constraints of non-holonomic motion specific to Ackermann steering models, conventions commonly used in vehicle dynamics. The adaptation of the non-holonomic ground vehicle kinematic model into VO tasks allows for more robust motion estimation. The study extends single feature n-linearities, typically used in traditional frame-based visual systems, to accommodate the quasi time-continuous dynamics captured through event cameras. The authors employ polynomial formulations via variable Taylor expansions to create a framework that accommodates dynamic, event-driven data streams.

The paper highlights an advanced approach of averaging across multiple event tracks using histogram voting, which improves the resilience of odometry results. This method shows improved performance on vehicle’s instantaneous rotational velocity estimation, demonstrated via extensive simulations and real-world data evaluations. Notably, the proposed method exhibits superior performance under challenging lighting conditions compared to traditional frame-based methods, confirming its potential for applications in environments where regular cameras are less effective.

Key contributions of the paper include:

  • Integration of a continuous-time incidence relation within the VO framework under the assumption of non-holonomic Ackermann motion and constant rotational velocity.
  • Utilization of Taylor series expansions to approximate trigonometric functions, generating n-linear constraints related to rotational velocity.
  • Deployment of rank minimization and histogram voting techniques for robust solution extraction and outlier rejection.

The implications of this research are both practical and theoretical. Practically, it provides a more reliable method for visual navigation in automated ground vehicles operating in challenging environments, potentially improving the efficacy and safety of autonomous vehicles. Theoretically, it broadens the application scope of event-based cameras within robotics, encouraging further study into non-holonomic constraints in diverse robotic and automotive applications.

Future research may explore the integration of more advanced sensor fusion techniques, such as incorporating inertial measurements, to further enhance the robustness and accuracy of estimation processes in motion prediction frameworks. Additionally, advancements in event-camera hardware and asynchronous signal processing could further bridge the performance gaps with traditional visual systems, unlocking new potential in real-time navigation and mapping tasks.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.