An effective and efficient green federated learning method for one-layer neural networks (2312.14528v1)
Abstract: Nowadays, machine learning algorithms continue to grow in complexity and require a substantial amount of computational resources and energy. For these reasons, there is a growing awareness of the development of new green algorithms and distributed AI can contribute to this. Federated learning (FL) is one of the most active research lines in machine learning, as it allows the training of collaborative models in a distributed way, an interesting option in many real-world environments, such as the Internet of Things, allowing the use of these models in edge computing devices. In this work, we present a FL method, based on a neural network without hidden layers, capable of generating a global collaborative model in a single training round, unlike traditional FL methods that require multiple rounds for convergence. This allows obtaining an effective and efficient model that simplifies the management of the training process. Moreover, this method preserve data privacy by design, a crucial aspect in current data protection regulations. We conducted experiments with large datasets and a large number of federated clients. Despite being based on a network model without hidden layers, it maintains in all cases competitive accuracy results compared to more complex state-of-the-art machine learning models. Furthermore, we show that the method performs equally well in both identically and non-identically distributed scenarios. Finally, it is an environmentally friendly algorithm as it allows significant energy savings during the training process compared to its centralized counterpart.
- Sustainable AI: Environmental implications, challenges and opportunities. CoRR, page 2111.00364, 2021.
- A first look into the carbon footprint of federated learning. arXiv, abs/2102.07627, 2021.
- P. Dhar. The carbon impact of artificial intelligence. Nature Machine Intelligence, 2:423–425, 2020.
- R. van der Meulen. What edge computing means for infrastructure and operations leaders. https://www.gartner.com/smarterwithgartner/what-edge-computing-means-for-infrastructure-and-operations-leaders, 2018. Accessed: 2023-12-21.
- Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, volume 54, pages 1273–1282. PMLR, 2017.
- Federated learning for mobile keyboard prediction. ArXiv, abs/1811.03604, 2018.
- Federated learning of predictive models from federated electronic health records. International Journal of Medical Informatics, 112:59–67, 2018.
- Federated AI for building AI solutions across multiple agencies. arXiv, abs/1809.10036, 2018.
- Quantifying the carbon emissions of machine learning. arXiv, abs/1910.09700, 2019.
- Towards the systematic reporting of the energy and carbon footprints of machine learning. The Journal of Machine Learning Research, 21(1):10039–10081, 2020.
- A framework for energy and carbon footprint analysis of distributed and federated edge learning. In 2021 IEEE 32nd Annual International Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC), pages 1564–1569, 2021.
- SCAFFOLD: Stochastic controlled averaging for federated learning. In Proceedings of the 37th International Conference on Machine Learning, volume 119, pages 5132–5143. PMLR, 2020.
- Mime: Mimicking centralized stochastic algorithms in federated learning. arXiv, abs/2008.03606, 2020.
- H. Alfauri and F. Esposito. A distributed consensus protocol for sustainable federated learning. In Proceedings of the 2021 IEEE Conference on Network Softwarization: Accelerating Network Softwarization in the Cognitive Age, NetSoft 2021, page 161 – 165, 2021.
- Fedheonn: Federated and homomorphically encrypted learning method for one-layer neural networks. Future Generation Computer Systems, 149:200–211, 2023.
- A new convex objective function for the supervised learning of single-layer neural networks. Pattern Recognition, 43(5):1984–1992, 2010.
- Regularized one-layer neural networks for distributed and incremental environments. In Advances in Computational Intelligence. IWANN 2021. Lecture Notes in Computer Science, volume 12862, pages 343–355. Springer International Publishing, 2021.
- A distributed and incremental SVD algorithm for agglomerative data analysis on large networks. SIAM Journal on Matrix Analysis and Applications, 37(4):1699–1718, 2016.
- D. Dua and C. Graff. UCI machine learning repository, 2017. University of California, Irvine, School of Information and Computer Sciences.
- Evolving large-scale data stream analytics based on scalable PANFIS. Knowledge-Based Systems, 166:186–197, 2019.
- Alterkawi Laila. Scaling Genetic Algorithms to Large Distributed Datasets. PhD thesis, University of Kent, 2022.
- Learning deep kernels for non-parametric two-sample tests. In Proceedings of the 37th International Conference on Machine Learning, pages 6316–6326, 2020.
- Higgs boson discovery using machine learning methods with pyspark. Procedia Computer Science, 170:1141–1146, 2020.
- Big data analytics based on PANFIS MapReduce. Procedia Computer Science, 144:140–152, 2018.
- Using Pyspark environment for solving a big data problem: Searching for supersymmetric particles. International Journal of Innovative Technology and Exploring Engineering, 9(7):541–546, 2020.
- A MapReduce solution for associative classification of big data. Information Sciences, 332:33–55, 2016.
- Rotation forest for big data. Information Fusion, 74:39–49, 2021.