On-device Self-supervised Learning of Visual Perception Tasks aboard Hardware-limited Nano-quadrotors (2403.04071v1)
Abstract: Sub-\SI{50}{\gram} nano-drones are gaining momentum in both academia and industry. Their most compelling applications rely on onboard deep learning models for perception despite severe hardware constraints (\ie sub-\SI{100}{\milli\watt} processor). When deployed in unknown environments not represented in the training data, these models often underperform due to domain shift. To cope with this fundamental problem, we propose, for the first time, on-device learning aboard nano-drones, where the first part of the in-field mission is dedicated to self-supervised fine-tuning of a pre-trained convolutional neural network (CNN). Leveraging a real-world vision-based regression task, we thoroughly explore performance-cost trade-offs of the fine-tuning phase along three axes: \textit{i}) dataset size (more data increases the regression performance but requires more memory and longer computation); \textit{ii}) methodologies (\eg fine-tuning all model parameters vs. only a subset); and \textit{iii}) self-supervision strategy. Our approach demonstrates an improvement in mean absolute error up to 30\% compared to the pre-trained baseline, requiring only \SI{22}{\second} fine-tuning on an ultra-low-power GWT GAP9 System-on-Chip. Addressing the domain shift problem via on-device learning aboard nano-drones not only marks a novel result for hardware-limited robots but lays the ground for more general advancements for the entire robotics community.
- J. Wang, C. Lan, C. Liu, Y. Ouyang, T. Qin, W. Lu, Y. Chen, W. Zeng, and P. S. Yu, “Generalizing to unseen domains: A survey on domain generalization,” IEEE Transactions on Knowledge and Data Engineering, vol. 35, no. 8, pp. 8052–8072, 2023.
- C. Heinze-Deml and N. Meinshausen, “Conditional variance penalties and domain shift robustness,” Machine Learning, vol. 110, no. 2, pp. 303–348, 2021.
- D. Palossi, N. Zimmerman, A. Burrello, F. Conti, H. Müller, L. M. Gambardella, L. Benini, A. Giusti, and J. Guzzi, “Fully onboard AI-powered human-drone pose estimation on ultralow-power autonomous flying nano-UAVs,” IEEE Internet of Things Journal, vol. 9, no. 3, pp. 1913–1929, 2021.
- L. Lamberti, V. Niculescu, M. Barciś, L. Bellone, E. Natalizio, L. Benini, and D. Palossi, “Tiny-PULP-Dronets: Squeezing neural networks for faster and lighter inference on multi-tasking autonomous nano-drones,” in 2022 IEEE 4th International Conference on Artificial Intelligence Circuits and Systems (AICAS), 2022, pp. 287–290.
- K. Kelchtermans and T. Tuytelaars, “RARA: Zero-shot Sim2Real visual navigation with following foreground cues,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2022, pp. 1704–1710.
- L. Lamberti, E. Cereda, G. Abbate, L. Bellone, V. J. K. Morinigo, M. Barciś, A. Barciś, A. Giusti, F. Conti, and D. Palossi, “A sim-to-real deep learning-based framework for autonomous nano-drone racing,” pp. 1899–1906, 2024.
- S. Jung, S. Cho, D. Lee, H. Lee, and D. H. Shim, “A direct visual servoing-based framework for the 2016 IROS Autonomous Drone Racing Challenge,” Journal of Field Robotics, vol. 35, no. 1, 2018.
- K. Mcguire, C. De Wagter, K. Tuyls, H. Kappen, and G. Croon, “Minimal Navigation Solution for a Swarm of Tiny Flying Robots to Explore an Unknown Environment,” Science Robotics, vol. 4, p. eaaw9710, Oct. 2019.
- S. Chen et al., “Towards specialized hardware for learning-based visual odometry on the edge,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2022, pp. 10 603–10 610.
- T. Lesort, V. Lomonaco, A. Stoian, D. Maltoni, D. Filliat, and N. Díaz-Rodríguez, “Continual learning for robotics: Definition, framework, learning strategies, opportunities and challenges,” Information Fusion, vol. 58, pp. 52–68, 2020.
- J. Eschmann, D. Albani, and G. Loianno, “RLtools: A fast, portable deep reinforcement learning library for continuous control,” 2023.
- J. Lin, L. Zhu, W.-M. Chen, W.-C. Wang, C. Gan, and S. Han, “On-device training under 256kB memory,” Advances in Neural Information Processing Systems, vol. 35, pp. 22 941–22 954, 2022.
- C. Profentzas, M. Almgren, and O. Landsiedel, “MiniLearn: On-device learning for low-power IoT devices,” in International Conference on Embedded Wireless Systems and Networks, 2022.
- M. Nava, L. M. Gambardella, and A. Giusti, “State-consistency loss for learning spatial perception tasks from partial labels,” IEEE Robotics and Automation Letters, vol. 6, no. 2, pp. 1112–1119, 2021.
- E. Cereda, D. Palossi, and A. Giusti, “Handling pitch variations for visual perception in MAVs: Synthetic augmentation and state fusion,” in 13th𝑡ℎ{}^{th}start_FLOATSUPERSCRIPT italic_t italic_h end_FLOATSUPERSCRIPT International Micro Air Vehicle Conference (IMAV), G. de Croon and C. D. Wagter, Eds., Sep 2022, pp. 59–65.
- A. Moldagalieva and W. Hönig, “Virtual omnidirectional perception for downwash prediction within a team of nano multirotors flying in close proximity,” 2023.
- L. Crupi, E. Cereda, A. Giusti, and D. Palossi, “Sim-to-real vision-depth fusion CNNs for robust pose estimation aboard autonomous nano-quadcopters,” in 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2023, pp. 7711–7717.
- M. Pavan, E. Ostrovan, A. Caltabiano, and M. Roveri, “TyBox: an automatic design and code-generation toolbox for TinyML incremental on-device learning,” ACM Transactions on Embedded Computing Systems, 2023.
- H. Ren, D. Anicic, and T. A. Runkler, “TinyOL: TinyML with online-learning on microcontrollers,” in 2021 International Joint Conference on Neural Networks (IJCNN). IEEE, 2021, pp. 1–8.
- D. P. Pau, P. K. Ambrose, and F. M. Aymone, “A quantitative review of automated neural search and on-device learning for tiny devices,” Chips, vol. 2, no. 2, p. 130–141, May 2023. [Online]. Available: http://dx.doi.org/10.3390/chips2020008
- C. D. Wagter, F. Paredes-Vallé, N. Sheth, and G. de Croon, “The sensing, state-estimation, and control behind the winning entry to the 2019 Artificial Intelligence Robotic Racing Competition,” Field Robotics, vol. 2, no. 1, pp. 1263–1290, mar 2022.
- R. J. Bouwmeester, F. Paredes-Vallés, and G. C. De Croon, “NanoFlowNet: Real-time dense optical flow on a nano quadcopter,” in 2023 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2023, pp. 1996–2003.
- K. Lamers, S. Tijmons, C. De Wagter, and G. de Croon, “Self-supervised monocular distance learning on a lightweight micro air vehicle,” in 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2016, pp. 1779–1784.
- D. Nadalini, M. Rusci, L. Benini, and F. Conti, “Reduced precision floating-point optimization for deep neural network on-device learning on microcontrollers,” Future Generation Computer Systems, vol. 149, pp. 212–226, 2023. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0167739X23002728
- Y. D. Kwon, R. Li, S. I. Venieris, J. Chauhan, N. D. Lane, and C. Mascolo, “TinyTrain: Deep neural network training at the extreme edge,” arXiv preprint arXiv:2307.09988, 2023.
- P. K. Mudrakarta, M. Sandler, A. Zhmoginov, and A. Howard, “K for the price of 1: Parameter efficient multi-task and transfer learning,” in International Conference on Learning Representations, 2019.
- H. Cai, C. Gan, L. Zhu, and S. Han, “TinyTL: Reduce memory, not parameters for efficient on-device learning,” Advances in Neural Information Processing Systems, vol. 33, pp. 11 285–11 297, 2020.
- D. Gandhi, L. Pinto, and A. Gupta, “Learning to fly by crashing,” in 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE Press, 2017, p. 3948–3955. [Online]. Available: https://doi.org/10.1109/IROS.2017.8206247
- A. Kouris and C.-S. Bouganis, “Learning to fly by myself: A self-supervised cnn-based approach for autonomous navigation,” in 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE Press, 2018, p. 1–9. [Online]. Available: https://doi.org/10.1109/IROS.2018.8594204
- M. Nava, A. Paolillo, J. Guzzi, L. M. Gambardella, and A. Giusti, “Learning visual localization of a quadrotor using its noise as self-supervision,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 2218–2225, 2022.
- I. Radosavovic, T. Xiao, S. James, P. Abbeel, J. Malik, and T. Darrell, “Real-world robot learning with masked visual pre-training,” CoRL, 2022.
- M. Nava, A. Paolillo, J. Guzzi, L. M. Gambardella, and A. Giusti, “Uncertainty-aware self-supervised learning of spatial perception tasks,” IEEE Robotics and Automation Letters, vol. 6, no. 4, pp. 6693–6700, 2021.
- G. Iyer, J. Krishna Murthy, G. Gupta, M. Krishna, and L. Paull, “Geometric consistency for self-supervised end-to-end visual odometry,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2018.
- J. Bian, Z. Li, N. Wang, H. Zhan, C. Shen, M.-M. Cheng, and I. Reid, “Unsupervised scale-consistent depth and ego-motion learning from monocular video,” in Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, Eds., vol. 32. Curran Associates, Inc., 2019.
- T. Zhou, M. Brown, N. Snavely, and D. G. Lowe, “Unsupervised learning of depth and ego-motion from video,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 1851–1858.
- C. Godard, O. Mac Aodha, M. Firman, and G. J. Brostow, “Digging into self-supervised monocular depth estimation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2019.
- P. Liu, M. Lyu, I. King, and J. Xu, “SelFlow: Self-supervised learning of optical flow,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019.
- A. Ghosh, H. Kumar, and P. S. Sastry, “Robust loss functions under label noise for deep neural networks,” in Proceedings of the AAAI conference on artificial intelligence, vol. 31, no. 1, 2017.
- Elia Cereda (11 papers)
- Manuele Rusci (19 papers)
- Alessandro Giusti (38 papers)
- Daniele Palossi (28 papers)