Amirkabir campus dataset: Real-world challenges and scenarios of Visual Inertial Odometry (VIO) for visually impaired people (2401.03604v1)
Abstract: Visual Inertial Odometry (VIO) algorithms estimate the accurate camera trajectory by using camera and Inertial Measurement Unit (IMU) sensors. The applications of VIO span a diverse range, including augmented reality and indoor navigation. VIO algorithms hold the potential to facilitate navigation for visually impaired individuals in both indoor and outdoor settings. Nevertheless, state-of-the-art VIO algorithms encounter substantial challenges in dynamic environments, particularly in densely populated corridors. Existing VIO datasets, e.g., ADVIO, typically fail to effectively exploit these challenges. In this paper, we introduce the Amirkabir campus dataset (AUT-VI) to address the mentioned problem and improve the navigation systems. AUT-VI is a novel and super-challenging dataset with 126 diverse sequences in 17 different locations. This dataset contains dynamic objects, challenging loop-closure/map-reuse, different lighting conditions, reflections, and sudden camera movements to cover all extreme navigation scenarios. Moreover, in support of ongoing development efforts, we have released the Android application for data capture to the public. This allows fellow researchers to easily capture their customized VIO dataset variations. In addition, we evaluate state-of-the-art Visual Inertial Odometry (VIO) and Visual Odometry (VO) methods on our dataset, emphasizing the essential need for this challenging dataset.
- Perception-aided visual-inertial integrated positioning in dynamic urban areas. In 2020 IEEE/ION Position, Location and Navigation Symposium (PLANS), pages 1563–1571. IEEE, 2020.
- Dynaslam: Tracking, mapping, and inpainting in dynamic scenes. IEEE Robotics and Automation Letters, 3(4):4076–4083, 2018.
- The málaga urban dataset: High-rate stereo and lidar in a realistic urban scenario. The International Journal of Robotics Research, 33(2):207–214, 2014.
- The euroc micro aerial vehicle datasets. The International Journal of Robotics Research, 35(10):1157–1163, 2016.
- Orb-slam3: An accurate open-source library for visual, visual–inertial, and multimap slam. IEEE Transactions on Robotics, 2021.
- The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3213–3223, 2016.
- Advio: An authentic dataset for visual-inertial odometry. In Proceedings of the European Conference on Computer Vision (ECCV), pages 419–434, 2018.
- A photometrically calibrated benchmark for monocular visual odometry. arXiv preprint arXiv:1607.02555, 2016.
- Unified temporal and spatial calibration for multi-sensor systems. In 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 1280–1286. IEEE, 2013.
- Urban@ cras dataset: Benchmarking of visual odometry and slam techniques. Robotics and Autonomous Systems, 109:59–67, 2018.
- Vision meets robotics: The kitti dataset. The International Journal of Robotics Research, 32(11):1231–1237, 2013.
- Complex urban dataset with multi-level sensors from highly diverse urban environments. The International Journal of Robotics Research, 38(6):642–657, 2019.
- The oxford multimotion dataset: Multiple se (3) motions with ground truth. IEEE Robotics and Automation Letters, 4(2):800–807, 2019.
- A benchmark for visual-inertial odometry systems employing onboard illumination. In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 5256–5263. IEEE, 2019.
- The zurich urban micro aerial vehicle dataset. The International Journal of Robotics Research, 36(3):269–273, 2017.
- Viode: A simulated dataset to address the challenges of visual-inertial odometry in dynamic environments. IEEE Robotics and Automation Letters, 6(2):1343–1350, 2021.
- Visual navigation features selection algorithm based on instance segmentation in dynamic environment. IEEE Access, 8:465–473, 2019.
- Penncosyvio: A challenging visual inertial odometry benchmark. In 2017 IEEE International Conference on Robotics and Automation (ICRA), pages 3847–3854. IEEE, 2017.
- Vins-mono: A robust and versatile monocular visual-inertial state estimator. IEEE Transactions on Robotics, 34(4):1004–1020, 2018.
- Rilp: Robust iranian license plate recognition designed for complex conditions. In 2020 International Conference on Machine Vision and Image Processing (MVIP), pages 1–7. IEEE, 2020.
- Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4938–4947, 2020.
- Slamantic-leveraging semantics to improve vslam in dynamic environments. In Proceedings of the IEEE International Conference on Computer Vision Workshops, pages 0–0, 2019.
- The tum vi benchmark for evaluating visual-inertial odometry. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 1680–1687. IEEE, 2018.
- The new college vision and laser data set. The International Journal of Robotics Research, 28(5):595–599, 2009.
- A benchmark for the evaluation of rgb-d slam systems. In 2012 IEEE/RSJ international conference on intelligent robots and systems, pages 573–580. IEEE, 2012.
- Visual-inertial mapping with non-linear factor recovery. IEEE Robotics and Automation Letters, 5(2):422–429, 2019.
- Dynamic-slam: Semantic monocular visual localization and mapping based on deep learning in dynamic environment. Robotics and Autonomous Systems, 117:1–16, 2019.
- A tutorial on quantitative trajectory evaluation for visual (-inertial) odometry. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 7244–7251. IEEE, 2018.
- The uma-vi dataset: Visual–inertial odometry in low-textured and dynamic illumination environments. The International Journal of Robotics Research, 39(9):1052–1060, 2020.