Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
129 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Aggregating Multiple Bio-Inspired Image Region Classifiers For Effective And Lightweight Visual Place Recognition (2312.12995v1)

Published 20 Dec 2023 in cs.CV

Abstract: Visual place recognition (VPR) enables autonomous systems to localize themselves within an environment using image information. While VPR techniques built upon a Convolutional Neural Network (CNN) backbone dominate state-of-the-art VPR performance, their high computational requirements make them unsuitable for platforms equipped with low-end hardware. Recently, a lightweight VPR system based on multiple bio-inspired classifiers, dubbed DrosoNets, has been proposed, achieving great computational efficiency at the cost of reduced absolute place retrieval performance. In this work, we propose a novel multi-DrosoNet localization system, dubbed RegionDrosoNet, with significantly improved VPR performance, while preserving a low-computational profile. Our approach relies on specializing distinct groups of DrosoNets on differently sliced partitions of the original image, increasing extrinsic model differentiation. Furthermore, we introduce a novel voting module to combine the outputs of all DrosoNets into the final place prediction which considers multiple top refence candidates from each DrosoNet. RegionDrosoNet outperforms other lightweight VPR techniques when dealing with both appearance changes and viewpoint variations. Moreover, it competes with computationally expensive methods on some benchmark datasets at a small fraction of their online inference time.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. S. Lowry, N. Sünderhauf, P. Newman, J. J. Leonard, D. Cox, P. Corke, and M. J. Milford, “Visual place recognition: A survey,” IEEE Transactions on Robotics, vol. 32, no. 1, pp. 1–19, 2015.
  2. F. Maffra, Z. Chen, and M. Chli, “tolerant place recognition combining 2d and 3d information for UAV navigation,” in 2018 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2018, pp. 2542–2549.
  3. A. Ranganathan, S. Matsumoto, and D. Ilstrup, “Towards illumination invariance for visual localization,” in 2013 IEEE International Conference on Robotics and Automation.   IEEE, 2013, pp. 3791–3798.
  4. T. Naseer, L. Spinello, W. Burgard, and C. Stachniss, “Robust visual robot localization across seasons using network flows,” in Twenty-eighth AAAI conference on artificial intelligence, 2014.
  5. A. Pronobis, B. Caputo, P. Jensfelt, and H. I. Christensen, “A discriminative approach to robust visual place recognition,” in 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.   IEEE, 2006, pp. 3829–3836.
  6. C.-C. Wang, C. Thorpe, S. Thrun, M. Hebert, and H. Durrant-Whyte, “Simultaneous localization, mapping and moving object tracking,” The International Journal of Robotics Research, vol. 26, no. 9, pp. 889–916, 2007.
  7. B. Ferrarini, M. Waheed, S. Waheed, S. Ehsan, M. Milford, and K. D. McDonald-Maier, “Visual place recognition for aerial robotics: Exploring accuracy-computation trade-off for local image descriptors,” in 2019 NASA/ESA Conference on Adaptive Hardware and Systems (AHS).   IEEE, 2019, pp. 103–108.
  8. Y. Hou, H. Zhang, and S. Zhou, “Convolutional neural network-based image representation for visual loop closure detection,” in 2015 IEEE International Conference on Information and Automation, 2015, pp. 2238–2245.
  9. B. Arcanjo, B. Ferrarini, M. Milford, K. D. McDonald-Maier, and S. Ehsan, “An efficient and scalable collection of fly-inspired voting units for visual place recognition in changing environments,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 2527–2534, 2022.
  10. T. A. Ofstad, C. S. Zuker, and M. B. Reiser, “Visual place learning in drosophila melanogaster,” Nature, vol. 474, no. 7350, pp. 204–207, 2011.
  11. N. Dalal and B. Triggs, “Histograms of oriented gradients for human detection,” in 2005 IEEE computer society conference on computer vision and pattern recognition (CVPR’05), vol. 1.   Ieee, 2005, pp. 886–893.
  12. C. McManus, B. Upcroft, and P. Newmann, “Scene signatures: Localised and point-less features for localisation,” in Proceedings of Robotics: Science and Systems, Berkeley, USA, July 2014.
  13. D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” International journal of computer vision, vol. 60, no. 2, pp. 91–110, 2004.
  14. H. Bay, A. Ess, T. Tuytelaars, and L. Van Gool, “Speeded-up robust features (SURF),” Computer vision and image understanding, vol. 110, no. 3, pp. 346–359, 2008.
  15. M. Zaffar, A. Khaliq, S. Ehsan, M. Milford, and K. McDonald-Maier, “Levelling the playing field: A comprehensive comparison of visual place recognition approaches under changing conditions,” ICRA 2019 Workshop on Database Generation and Benchmarking of SLAM Algorithms for Robotics and VR/AR, 2019.
  16. M. Zaffar, A. Khaliq, S. Ehsan, M. Milford, K. Alexis, and K. McDonald-Maier, “Are state-of-the-art visual place recognition techniques any good for aerial robotics?” IEEE ICRA 2019 Workshop on Aerial Robotics, 2019.
  17. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Advances in neural information processing systems, vol. 25, pp. 1097–1105, 2012.
  18. B. Zhou, A. Lapedriza, A. Khosla, A. Oliva, and A. Torralba, “Places: A 10 million image database for scene recognition,” IEEE transactions on pattern analysis and machine intelligence, vol. 40, no. 6, pp. 1452–1464, 2017.
  19. Z. Chen, A. Jacobson, N. Sünderhauf, B. Upcroft, L. Liu, C. Shen, I. Reid, and M. Milford, “Deep learning features at scale for visual place recognition,” in 2017 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2017, pp. 3223–3230.
  20. R. Arandjelovic, P. Gronat, A. Torii, T. Pajdla, and J. Sivic, “NetVLAD: CNN architecture for weakly supervised place recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 5297–5307.
  21. S. Hausler, S. Garg, M. Xu, M. Milford, and T. Fischer, “Patch-netvlad: Multi-scale fusion of locally-global descriptors for place recognition,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2021, pp. 14 141–14 152.
  22. M. Zaffar, S. Ehsan, M. Milford, and K. McDonald-Maier, “CoHOG: A light-weight, compute-efficient, and training-free visual place recognition technique for changing environments,” IEEE Robotics and Automation Letters, vol. 5, no. 2, pp. 1835–1842, 2020.
  23. N. Merrill and G. Huang, “Lightweight unsupervised deep loop closure,” arXiv preprint arXiv:1805.07703, 2018.
  24. A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and H. Adam, “Mobilenets: Efficient convolutional neural networks for mobile vision applications,” arXiv preprint arXiv:1704.04861, 2017.
  25. J. Yang, X. Shen, J. Xing, X. Tian, H. Li, B. Deng, J. Huang, and X.-s. Hua, “Quantization networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 7308–7316.
  26. B. Ferrarini, M. Milford, K. D. McDonald-Maier, and S. Ehsan, “Highly-efficient binary neural networks for visual place recognition,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2022, pp. 5493–5500.
  27. A. Cope, C. Sabo, E. Yavuz, K. Gurney, J. Marshall, T. Nowotny, and E. Vasilaki, “The green brain project–developing a neuromimetic robotic honeybee,” in Conference on Biomimetic and Biohybrid Systems.   Springer, 2013, pp. 362–363.
  28. A. Narendra, S. Gourmaud, and J. Zeil, “Mapping the navigational knowledge of individually foraging ants, myrmecia croslandi,” Proceedings of the Royal Society B: Biological Sciences, vol. 280, no. 1765, p. 20130683, 2013.
  29. M. J. Milford, G. F. Wyeth, and D. Prasser, “RatSLAM: a hippocampal model for simultaneous localization and mapping,” in IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA’04. 2004, vol. 1.   IEEE, 2004, pp. 403–408.
  30. M. Chancán, L. Hernandez-Nunez, A. Narendra, A. B. Barron, and M. Milford, “A hybrid compact neural architecture for visual place recognition,” IEEE Robotics and Automation Letters, vol. 5, no. 2, pp. 993–1000, 2020.
  31. S. Dasgupta, C. F. Stevens, and S. Navlakha, “A neural algorithm for a fundamental computing problem,” Science, vol. 358, no. 6364, pp. 793–796, 2017.
  32. N. Sünderhauf, P. Neubert, and P. Protzel, “Are we there yet? challenging SeqSLAM on a 3000 km journey across all four seasons,” in Proc. of Workshop on Long-Term Autonomy, IEEE International Conference on Robotics and Automation (ICRA).   Citeseer, 2013, p. 2013.
  33. N. Sünderhauf, S. Shirazi, F. Dayoub, B. Upcroft, and M. Milford, “On the performance of convnet features for place recognition,” in 2015 IEEE/RSJ international conference on intelligent robots and systems (IROS).   IEEE, 2015, pp. 4297–4304.
  34. A. J. Glover, W. P. Maddern, M. J. Milford, and G. F. Wyeth, “Fab-map+ ratslam: Appearance-based slam for multiple times of day,” in 2010 IEEE international conference on robotics and automation.   IEEE, 2010, pp. 3507–3512.
  35. Z. Chen, F. Maffra, I. Sa, and M. Chli, “Only look once, mining distinctive landmarks from convnet for visual place recognition,” in 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2017, pp. 9–16.
  36. F. Maffra, L. Teixeira, Z. Chen, and M. Chli, “Real-time wide-baseline place recognition using depth completion,” IEEE Robotics and Automation Letters, 2019.
  37. D. M. Powers, “Evaluation: from precision, recall and f-measure to roc, informedness, markedness and correlation,” Journal of Machine Learning Technologies, vol. 2, no. 1, pp. 37–63, 2011.
  38. M. Zaffar, S. Garg, M. Milford, J. Kooij, D. Flynn, K. McDonald-Maier, and S. Ehsan, “Vpr-bench: An open-source visual place recognition evaluation framework with quantifiable viewpoint and appearance change,” International Journal of Computer Vision, pp. 1–39, 2021.
  39. J. Davis and M. Goadrich, “The relationship between Precision-Recall and ROC curves,” in Proceedings of the 23rd international conference on Machine learning, 2006, pp. 233–240.
  40. S. Schubert, P. Neubert, S. Garg, M. Milford, and T. Fischer, “Visual place recognition: A tutorial,” 2023.
  41. D. Bai, C. Wang, B. Zhang, X. Yi, and X. Yang, “Sequence searching with cnn features for robust and fast visual place recognition,” Computers & Graphics, vol. 70, pp. 270–280, 2018.
  42. B. Ferrarini, M. Waheed, S. Waheed, S. Ehsan, M. J. Milford, and K. D. McDonald-Maier, “Exploring performance bounds of visual place recognition using extended precision,” IEEE Robotics and Automation Letters, vol. 5, no. 2, pp. 1688–1695, April 2020.
  43. D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” 3rd International Conference for Learning Representations, San Diego, 2015, 2015.

Summary

  • The paper introduces RegionDrosoNet, a bio-inspired VPR system that aggregates specialized DrosoNets to enhance place recognition on resource-constrained robots.
  • The method divides images into non-overlapping regions and applies grid-specific DrosoNets, improving recognition across seasonal and viewpoint variations.
  • Benchmark tests show that RegionDrosoNet achieves real-time performance comparable to heavier CNN-based systems while maintaining high efficiency.

Introduction

Visual place recognition (VPR) has become an essential technology for mobile robotics, allowing robots to determine their location in an environment based solely on visual inputs, typically camera images. Traditional VPR systems often rely on Convolutional Neural Networks (CNNs), but such models have high computational demands, which can be impractical for robots equipped with limited hardware capabilities.

Lightweight Approach to VPR

A novel VPR system known as RegionDrosoNet offers a balance between computational efficiency and VPR performance. Inspired by biological systems, specifically the fruit fly's olfactory circuitry, the system introduces a group of compact neural network models called DrosoNets. These models are trained to specialize in recognizing different subdivisions of an image, thereby diversifying their ability to match locations amid varying appearances and perspectives.

System Design and Functionality

RegionDrosoNet divides input images into multiple non-overlapping regions using different grid configurations, ensuring heterogeneous coverage. Each region feeds into a specialized group of DrosoNets, trained exclusively on that section of the image during the training phase. During operation, these DrosoNets generate multiple place confidence scores, which are then synthesized using a novel voting module that considers the top reference candidates from each. This approach allows RegionDrosoNet to reach consensus on the most accurate location match, increasing the reliability of VPR even when individual DrosoNets fail to recognize the correct place.

Performance and Evaluation

RegionDrosoNet's performance was evaluated using several benchmark datasets that test against seasonal variation, viewpoint changes, and different weather conditions. The system demonstrates a remarkable ability to outperform other lightweight VPR techniques, as well as bear comparison with more computationally demanding methods, all while maintaining a rapid inference time suitable for real-time applications.

Implications and Future Work

The implications of RegionDrosoNet's design are profound for the development of autonomous systems. By offering a VPR system that doesn't compromise on speed or accuracy, robots can potentially operate in various environments for extended periods without the need for high-end computational hardware. Future research directions could explore eliminating the need for environment-specific training, broadening the system's applicability across diverse deployment scenarios.

X Twitter Logo Streamline Icon: https://streamlinehq.com