Enabling On-device Continual Learning with Binary Neural Networks (2401.09916v1)
Abstract: On-device learning remains a formidable challenge, especially when dealing with resource-constrained devices that have limited computational capabilities. This challenge is primarily rooted in two key issues: first, the memory available on embedded devices is typically insufficient to accommodate the memory-intensive back-propagation algorithm, which often relies on floating-point precision. Second, the development of learning algorithms on models with extreme quantization levels, such as Binary Neural Networks (BNNs), is critical due to the drastic reduction in bit representation. In this study, we propose a solution that combines recent advancements in the field of Continual Learning (CL) and Binary Neural Networks to enable on-device training while maintaining competitive performance. Specifically, our approach leverages binary latent replay (LR) activations and a novel quantization scheme that significantly reduces the number of bits required for gradient computation. The experimental validation demonstrates a significant accuracy improvement in combination with a noticeable reduction in memory requirement, confirming the suitability of our approach in expanding the practical applications of deep learning in real-world scenarios.
- Esraa Mohamed “The relation of artificial intelligence with internet of things: A survey” In Journal of Cybersecurity and Information Management 1.1, 2020, pp. 30–24
- “A comprehensive survey of the Internet of Things (IoT) and AI-based smart healthcare” In IEEE Access 9 IEEE, 2020, pp. 3660–3678
- “Continual lifelong learning with neural networks: A review” In Neural networks 113 Elsevier, 2019
- “Class-incremental learning: survey and performance evaluation on image classification” In IEEE Transactions on Pattern Analysis and Machine Intelligence 45.5 IEEE, 2022, pp. 5513–5533
- “Overcoming catastrophic forgetting in neural networks” In Proceedings of the national academy of sciences 114.13 National Acad Sciences, 2017, pp. 3521–3526
- “Benchmarking tinyml systems: Challenges and direction” In arXiv preprint arXiv:2003.04821, 2020
- “Binarized neural networks: Training deep neural networks with weights and activations constrained to+ 1 or-1” In arXiv preprint arXiv:1602.02830, 2016
- “Xnor-net: Imagenet classification using binary convolutional neural networks” In European conference on computer vision, 2016, pp. 525–542 Springer
- “Binary neural networks: A survey” In Pattern Recognition 105 Elsevier, 2020, pp. 107281
- Lorenzo Vorabbi, Davide Maltoni and Stefano Santi “On-Device Learning with Binary Neural Networks” In arXiv preprint arXiv:2308.15308, 2023
- Vincenzo Lomonaco, Davide Maltoni and Lorenzo Pellegrini “Rehearsal-Free Continual Learning over Small Non-IID Batches.” In CVPR Workshops 1.2, 2020, pp. 3
- Gabriele Graffieti, Guido Borghi and Davide Maltoni “Continual learning in real-life applications” In IEEE Robotics and Automation Letters 7.3 IEEE, 2022, pp. 6195–6202
- “Latent replay for real-time continual learning” In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2020, pp. 10203–10209 IEEE
- “Latent weights do not exist: Rethinking binarized neural network optimization” In Advances in neural information processing systems 32, 2019
- “Tinytl: Reduce memory, not parameters for efficient on-device learning” In Advances in Neural Information Processing Systems 33, 2020, pp. 11285–11297
- “On-device training under 256kb memory” In Advances in Neural Information Processing Systems 35, 2022, pp. 22941–22954
- Haoyu Ren, Darko Anicic and Thomas A Runkler “Tinyol: Tinyml with online-learning on microcontrollers” In 2021 International Joint Conference on Neural Networks (IJCNN), 2021, pp. 1–8 IEEE
- “A tinyml platform for on-device continual learning with quantized latent replays” In IEEE Journal on Emerging and Selected Topics in Circuits and Systems 11.4 IEEE, 2021, pp. 789–802
- “PULP-TrainLib: Enabling on-device training for RISC-V multi-core MCUs through performance-driven Autotuning” In International Conference on Embedded Computer Systems, 2022, pp. 200–216 Springer
- “Reduced Precision Floating-Point Optimization for Deep Neural Network On-Device Learning on MicroControllers” In arXiv preprint arXiv:2305.19167, 2023
- “Larq compute engine: Design, benchmark and deploy state-of-the-art binarized neural networks” In Proceedings of Machine Learning and Systems 3, 2021, pp. 680–695
- L. Vorabbi, D. Maltoni and S. Santi “Optimizing data-flow in Binary Neural Networks”, 2023 arXiv:2304.00952 [cs.LG]
- “Batch normalization: Accelerating deep network training by reducing internal covariate shift” In International conference on machine learning, 2015, pp. 448–456 pmlr
- Jeffrey S Vitter “Random sampling with a reservoir” In ACM Transactions on Mathematical Software (TOMS) 11.1 ACM New York, NY, USA, 1985, pp. 37–57
- “Quantization and training of neural networks for efficient integer-arithmetic-only inference” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 2704–2713
- Benoit Jacob, P Warden and ME Guney “gemmlowp: a small self-contained low-precision GEMM library.(2017)” In URL https://github. com/google/gemmlowp, 2017
- “Bi-real net: Enhancing the performance of 1-bit cnns with improved representational capability and advanced training algorithm” In Proceedings of the European conference on computer vision (ECCV), 2018, pp. 722–737
- “Reactnet: Towards precise binary neural network with generalized activation functions” In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XIV 16, 2020, pp. 143–159 Springer
- “Deep learning with limited numerical precision” In International conference on machine learning, 2015, pp. 1737–1746 PMLR
- “Mixed precision training of convolutional neural networks using integer operations” In arXiv preprint arXiv:1802.00930, 2018
- “Scalable methods for 8-bit training of neural networks” In Advances in neural information processing systems 31, 2018
- “Core50: a new dataset and benchmark for continuous object recognition” In Conference on robot learning, 2017, pp. 17–26 PMLR
- “Learning multiple layers of features from tiny images” Toronto, ON, Canada, 2009
- “Tiny imagenet visual recognition challenge” In CS 231N 7.7, 2015, pp. 3
- “Back to simplicity: How to train accurate bnns from scratch?” In arXiv preprint arXiv:1906.08637, 2019
- “Training binary neural networks with real-to-binary convolutions” In arXiv preprint arXiv:2003.11535, 2020
- “Mobilenets: Efficient convolutional neural networks for mobile vision applications” In arXiv preprint arXiv:1704.04861, 2017
- Diederik P Kingma and Jimmy Ba “Adam: A method for stochastic optimization” In arXiv preprint arXiv:1412.6980, 2014
- Lorenzo Vorabbi (3 papers)
- Davide Maltoni (33 papers)
- Guido Borghi (26 papers)
- Stefano Santi (3 papers)