FedADMM-InSa: An Inexact and Self-Adaptive ADMM for Federated Learning (2402.13989v3)
Abstract: Federated learning (FL) is a promising framework for learning from distributed data while maintaining privacy. The development of efficient FL algorithms encounters various challenges, including heterogeneous data and systems, limited communication capacities, and constrained local computational resources. Recently developed FedADMM methods show great resilience to both data and system heterogeneity. However, they still suffer from performance deterioration if the hyperparameters are not carefully tuned. To address this issue, we propose an inexact and self-adaptive FedADMM algorithm, termed FedADMM-InSa. First, we design an inexactness criterion for the clients' local updates to eliminate the need for empirically setting the local training accuracy. This inexactness criterion can be assessed by each client independently based on its unique condition, thereby reducing the local computational cost and mitigating the undesirable straggle effect. The convergence of the resulting inexact ADMM is proved under the assumption of strongly convex loss functions. Additionally, we present a self-adaptive scheme that dynamically adjusts each client's penalty parameter, enhancing algorithm robustness by mitigating the need for empirical penalty parameter choices for each client. Extensive numerical experiments on both synthetic and real-world datasets are conducted. As validated by some numerical tests, our proposed algorithm can reduce the clients' local computational load significantly and also accelerate the learning process compared to the vanilla FedADMM.
- “Federated Learning Based on Dynamic Regularization” In arXiv preprint arXiv:2111.04263, 2021 DOI: 10.48550/arXiv.2111.04263
- Mostafa D. Awheda and Howard M. Schwartz “Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms” In Artificial Intelligence Review 45.3, 2016, pp. 299–332 DOI: 10.1007/s10462-015-9447-5
- “Towards Federated Learning at Scale: System Design” In Proceedings of machine learning and systems 1, 2019, pp. 374–388
- “Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers” In Foundations and Trends® in Machine learning 3.1 Now Publishers, Inc., 2011, pp. 1–122 DOI: 10.1561/2200000016
- “Exponential Moving Average Normalization for Self-Supervised and Semi-Supervised Learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 194–203
- Canh T. Dinh, Nguyen Tran and Josh Nguyen “Personalized Federated Learning with Moreau Envelopes” In Advances in Neural Information Processing Systems 33 Curran Associates, Inc., 2020, pp. 21394–21405
- “Improved Gradient Inversion Attacks and Defenses in Federated Learning” In IEEE Transactions on Big Data, 2023, pp. 1–13 DOI: 10.1109/TBDATA.2023.3239116
- Roland Glowinski “On Alternating Direction Methods of Multipliers: A Historical Perspective” In Modeling, Simulation and Optimization for Science and Technology 34 Dordrecht: Springer Netherlands, 2014, pp. 59–82 DOI: 10.1007/978-94-017-9054-3˙4
- “Sur l’approximation, Par Éléments Finis d’ordre Un, et La Résolution, Par Pénalisation-Dualité d’une Classe de Problèmes de Dirichlet Non Linéaires” In Revue française d’automatique, informatique, recherche opérationnelle. Analyse numérique 9.R2 EDP Sciences, 1975, pp. 41–76 DOI: 10.1051/m2an/197509R200411
- Roland Glowinski, Yongcun Song and Xiaoming Yuan “An ADMM Numerical Approach to Linear Parabolic State Constrained Optimal Control Problems” In Numerische Mathematik 144.4, 2020, pp. 931–966 DOI: 10.1007/s00211-020-01104-4
- “Application of the Alternating Direction Method of Multipliers to Control Constrained Parabolic Optimal Control Problems and Beyond” In Annals of Applied Mathematics 38, 2022 DOI: 10.4208/aam.OA-2022-0004
- Tom Goldstein, Min Li and Xiaoming Yuan “Adaptive Primal-Dual Splitting Methods for Statistical Learning and Image Processing” In Advances in Neural Information Processing Systems 28 Curran Associates, Inc., 2015
- Yonghai Gong, Yichuan Li and Nikolaos M. Freris “FedADMM: A Robust Federated Deep Learning Framework with Adaptivity to System Heterogeneity” In 2022 IEEE 38th International Conference on Data Engineering (ICDE), 2022, pp. 2575–2587 DOI: 10.1109/ICDE53745.2022.00238
- Bingsheng He, Hwho Yang and Swho Wang “Alternating Direction Method with Self-Adaptive Penalty Parameters for Monotone Variational Inequalities” In Journal of Optimization Theory and Applications 106.2, 2000, pp. 337–356 DOI: 10.1023/A:1004603514434
- “A Class of ADMM-based Algorithms for Three-Block Separable Convex Programming” In Computational Optimization and Applications 70.3, 2018, pp. 791–826 DOI: 10.1007/s10589-018-9994-1
- “Deep Residual Learning for Image Recognition” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778
- Magnus R. Hestenes “Multiplier and Gradient Methods” In Journal of optimization theory and applications 4.5 Springer, 1969, pp. 303–320 DOI: 10.1007/BF00927673
- “Advances and Open Problems in Federated Learning” In Foundations and Trends® in Machine Learning 14.1–2 Now Publishers, Inc., 2021, pp. 1–210 DOI: 10.1561/2200000083
- “SCAFFOLD: Stochastic Controlled Averaging for Federated Learning” In Proceedings of the 37th International Conference on Machine Learning PMLR, 2020, pp. 5132–5143
- “Gradient-Based Learning Applied to Document Recognition” In Proceedings of the IEEE 86.11 Ieee, 1998, pp. 2278–2324 DOI: 10.1109/5.726791
- “A Survey on Federated Learning Systems: Vision, Hype and Reality for Data Privacy and Protection” In IEEE Transactions on Knowledge and Data Engineering, 2021, pp. 1–1 DOI: 10.1109/TKDE.2021.3124599
- “Federated Learning: Challenges, Methods, and Future Directions” In IEEE Signal Processing Magazine 37.3, 2020, pp. 50–60 DOI: 10.1109/MSP.2020.2975749
- “Federated Optimization in Heterogeneous Networks” In Proceedings of Machine Learning and Systems 2, 2020, pp. 429–450
- “On the Convergence of FedAvg on Non-IID Data” In International Conference on Learning Representations, 2019
- Dong C. Liu and Jorge Nocedal “On the Limited Memory BFGS Method for Large Scale Optimization” In Mathematical Programming 45.1, 1989, pp. 503–528 DOI: 10.1007/BF01589116
- “Communication-Efficient Learning of Deep Networks from Decentralized Data” In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics PMLR, 2017, pp. 1273–1282
- Michael JD Powell “A Method for Nonlinear Constraints in Minimization Problems” In Optimization Academic Press, 1969, pp. 283–298
- Changkyu Song, Sejong Yoon and Vladimir Pavlovic “Fast ADMM Algorithm for Distributed Optimization with Adaptive Penalty” In Proceedings of the AAAI Conference on Artificial Intelligence 30.1, 2016 DOI: 10.1609/aaai.v30i1.10069
- Yongcun Song, Xiaoming Yuan and Hangrui Yue “The ADMM-PINNs Algorithmic Framework for Nonsmooth PDE-Constrained Optimization: A Deep Learning Approach” In arXiv preprint arXiv:2302.08309, 2023 DOI: 10.48550/arXiv.2302.08309
- “Towards Personalized Federated Learning” In IEEE Transactions on Neural Networks and Learning Systems 34.12, 2023, pp. 9587–9603 DOI: 10.1109/TNNLS.2022.3160699
- Han Wang, Siddartha Marella and James Anderson “FedADMM: A Federated Primal-Dual Algorithm Allowing Partial Participation” In 2022 IEEE 61st Conference on Decision and Control (CDC), 2022, pp. 287–294 DOI: 10.1109/CDC51059.2022.9992745
- “Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization” In Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS’20 Red Hook, NY, USA: Curran Associates Inc., 2020, pp. 7611–7623
- “Beyond ADMM: A Unified Client-Variance-Reduced Adaptive Federated Learning Framework” In Proceedings of the AAAI Conference on Artificial Intelligence 37.8, 2023, pp. 10175–10183 DOI: 10.1609/aaai.v37i8.26212
- Ziqi Wang, Yongcun Song and Enrique Zuazua “Approximate and Weighted Data Reconstruction Attack in Federated Learning” In arXiv preprint arXiv:2308.06822, 2023 DOI: 10.48550/arXiv.2308.06822
- Di Xiao, Jinkun Li and Min Li “Privacy-Preserving Federated Compressed Learning Against Data Reconstruction Attacks Based on Secure Data” In Neural Information Processing, Communications in Computer and Information Science Singapore: Springer Nature, 2024, pp. 325–339 DOI: 10.1007/978-981-99-8184-7˙25
- “ADMM without a Fixed Penalty Parameter: Faster Convergence with New Adaptive Penalization” In Advances in Neural Information Processing Systems 30 Curran Associates, Inc., 2017
- “Reveal Your Images: Gradient Leakage Attack against Unbiased Sampling-Based Secure Aggregation” In IEEE Transactions on Knowledge and Data Engineering, 2023, pp. 1–14 DOI: 10.1109/TKDE.2023.3271432
- “Implementing the Alternating Direction Method of Multipliers for Big Datasets: A Case Study of Least Absolute Shrinkage and Selection Operator” In SIAM Journal on Scientific Computing 40.5, 2018, pp. A3121–A3156 DOI: 10.1137/17M1146567
- “An Alternating Direction Method of Multipliers for Elliptic Equation Constrained Optimization Problem” In Science China Mathematics 60 Springer, 2017, pp. 361–378 DOI: 10.1007/s11425-015-0522-3
- “Federated Learning for the Internet of Things: Applications, Challenges, and Opportunities” In IEEE Internet of Things Magazine 5.1, 2022, pp. 24–29 DOI: 10.1109/IOTM.004.2100182
- “FedPD: A Federated Learning Framework With Adaptivity to Non-IID Data” In IEEE Transactions on Signal Processing 69, 2021, pp. 6055–6070 DOI: 10.1109/TSP.2021.3115952
- Shenglong Zhou and Geoffrey Ye Li “Federated Learning Via Inexact ADMM” In IEEE Transactions on Pattern Analysis and Machine Intelligence 45.8, 2023, pp. 9699–9708 DOI: 10.1109/TPAMI.2023.3243080
- “Parallelized Stochastic Gradient Descent” In Advances in Neural Information Processing Systems 23 Curran Associates, Inc., 2010
- Yongcun Song (12 papers)
- Ziqi Wang (93 papers)
- Enrique Zuazua (102 papers)