Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

FedADMM-InSa: An Inexact and Self-Adaptive ADMM for Federated Learning (2402.13989v3)

Published 21 Feb 2024 in cs.LG, cs.CR, cs.DC, and math.OC

Abstract: Federated learning (FL) is a promising framework for learning from distributed data while maintaining privacy. The development of efficient FL algorithms encounters various challenges, including heterogeneous data and systems, limited communication capacities, and constrained local computational resources. Recently developed FedADMM methods show great resilience to both data and system heterogeneity. However, they still suffer from performance deterioration if the hyperparameters are not carefully tuned. To address this issue, we propose an inexact and self-adaptive FedADMM algorithm, termed FedADMM-InSa. First, we design an inexactness criterion for the clients' local updates to eliminate the need for empirically setting the local training accuracy. This inexactness criterion can be assessed by each client independently based on its unique condition, thereby reducing the local computational cost and mitigating the undesirable straggle effect. The convergence of the resulting inexact ADMM is proved under the assumption of strongly convex loss functions. Additionally, we present a self-adaptive scheme that dynamically adjusts each client's penalty parameter, enhancing algorithm robustness by mitigating the need for empirical penalty parameter choices for each client. Extensive numerical experiments on both synthetic and real-world datasets are conducted. As validated by some numerical tests, our proposed algorithm can reduce the clients' local computational load significantly and also accelerate the learning process compared to the vanilla FedADMM.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. “Federated Learning Based on Dynamic Regularization” In arXiv preprint arXiv:2111.04263, 2021 DOI: 10.48550/arXiv.2111.04263
  2. Mostafa D. Awheda and Howard M. Schwartz “Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms” In Artificial Intelligence Review 45.3, 2016, pp. 299–332 DOI: 10.1007/s10462-015-9447-5
  3. “Towards Federated Learning at Scale: System Design” In Proceedings of machine learning and systems 1, 2019, pp. 374–388
  4. “Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers” In Foundations and Trends® in Machine learning 3.1 Now Publishers, Inc., 2011, pp. 1–122 DOI: 10.1561/2200000016
  5. “Exponential Moving Average Normalization for Self-Supervised and Semi-Supervised Learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 194–203
  6. Canh T. Dinh, Nguyen Tran and Josh Nguyen “Personalized Federated Learning with Moreau Envelopes” In Advances in Neural Information Processing Systems 33 Curran Associates, Inc., 2020, pp. 21394–21405
  7. “Improved Gradient Inversion Attacks and Defenses in Federated Learning” In IEEE Transactions on Big Data, 2023, pp. 1–13 DOI: 10.1109/TBDATA.2023.3239116
  8. Roland Glowinski “On Alternating Direction Methods of Multipliers: A Historical Perspective” In Modeling, Simulation and Optimization for Science and Technology 34 Dordrecht: Springer Netherlands, 2014, pp. 59–82 DOI: 10.1007/978-94-017-9054-3˙4
  9. “Sur l’approximation, Par Éléments Finis d’ordre Un, et La Résolution, Par Pénalisation-Dualité d’une Classe de Problèmes de Dirichlet Non Linéaires” In Revue française d’automatique, informatique, recherche opérationnelle. Analyse numérique 9.R2 EDP Sciences, 1975, pp. 41–76 DOI: 10.1051/m2an/197509R200411
  10. Roland Glowinski, Yongcun Song and Xiaoming Yuan “An ADMM Numerical Approach to Linear Parabolic State Constrained Optimal Control Problems” In Numerische Mathematik 144.4, 2020, pp. 931–966 DOI: 10.1007/s00211-020-01104-4
  11. “Application of the Alternating Direction Method of Multipliers to Control Constrained Parabolic Optimal Control Problems and Beyond” In Annals of Applied Mathematics 38, 2022 DOI: 10.4208/aam.OA-2022-0004
  12. Tom Goldstein, Min Li and Xiaoming Yuan “Adaptive Primal-Dual Splitting Methods for Statistical Learning and Image Processing” In Advances in Neural Information Processing Systems 28 Curran Associates, Inc., 2015
  13. Yonghai Gong, Yichuan Li and Nikolaos M. Freris “FedADMM: A Robust Federated Deep Learning Framework with Adaptivity to System Heterogeneity” In 2022 IEEE 38th International Conference on Data Engineering (ICDE), 2022, pp. 2575–2587 DOI: 10.1109/ICDE53745.2022.00238
  14. Bingsheng He, Hwho Yang and Swho Wang “Alternating Direction Method with Self-Adaptive Penalty Parameters for Monotone Variational Inequalities” In Journal of Optimization Theory and Applications 106.2, 2000, pp. 337–356 DOI: 10.1023/A:1004603514434
  15. “A Class of ADMM-based Algorithms for Three-Block Separable Convex Programming” In Computational Optimization and Applications 70.3, 2018, pp. 791–826 DOI: 10.1007/s10589-018-9994-1
  16. “Deep Residual Learning for Image Recognition” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778
  17. Magnus R. Hestenes “Multiplier and Gradient Methods” In Journal of optimization theory and applications 4.5 Springer, 1969, pp. 303–320 DOI: 10.1007/BF00927673
  18. “Advances and Open Problems in Federated Learning” In Foundations and Trends® in Machine Learning 14.1–2 Now Publishers, Inc., 2021, pp. 1–210 DOI: 10.1561/2200000083
  19. “SCAFFOLD: Stochastic Controlled Averaging for Federated Learning” In Proceedings of the 37th International Conference on Machine Learning PMLR, 2020, pp. 5132–5143
  20. “Gradient-Based Learning Applied to Document Recognition” In Proceedings of the IEEE 86.11 Ieee, 1998, pp. 2278–2324 DOI: 10.1109/5.726791
  21. “A Survey on Federated Learning Systems: Vision, Hype and Reality for Data Privacy and Protection” In IEEE Transactions on Knowledge and Data Engineering, 2021, pp. 1–1 DOI: 10.1109/TKDE.2021.3124599
  22. “Federated Learning: Challenges, Methods, and Future Directions” In IEEE Signal Processing Magazine 37.3, 2020, pp. 50–60 DOI: 10.1109/MSP.2020.2975749
  23. “Federated Optimization in Heterogeneous Networks” In Proceedings of Machine Learning and Systems 2, 2020, pp. 429–450
  24. “On the Convergence of FedAvg on Non-IID Data” In International Conference on Learning Representations, 2019
  25. Dong C. Liu and Jorge Nocedal “On the Limited Memory BFGS Method for Large Scale Optimization” In Mathematical Programming 45.1, 1989, pp. 503–528 DOI: 10.1007/BF01589116
  26. “Communication-Efficient Learning of Deep Networks from Decentralized Data” In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics PMLR, 2017, pp. 1273–1282
  27. Michael JD Powell “A Method for Nonlinear Constraints in Minimization Problems” In Optimization Academic Press, 1969, pp. 283–298
  28. Changkyu Song, Sejong Yoon and Vladimir Pavlovic “Fast ADMM Algorithm for Distributed Optimization with Adaptive Penalty” In Proceedings of the AAAI Conference on Artificial Intelligence 30.1, 2016 DOI: 10.1609/aaai.v30i1.10069
  29. Yongcun Song, Xiaoming Yuan and Hangrui Yue “The ADMM-PINNs Algorithmic Framework for Nonsmooth PDE-Constrained Optimization: A Deep Learning Approach” In arXiv preprint arXiv:2302.08309, 2023 DOI: 10.48550/arXiv.2302.08309
  30. “Towards Personalized Federated Learning” In IEEE Transactions on Neural Networks and Learning Systems 34.12, 2023, pp. 9587–9603 DOI: 10.1109/TNNLS.2022.3160699
  31. Han Wang, Siddartha Marella and James Anderson “FedADMM: A Federated Primal-Dual Algorithm Allowing Partial Participation” In 2022 IEEE 61st Conference on Decision and Control (CDC), 2022, pp. 287–294 DOI: 10.1109/CDC51059.2022.9992745
  32. “Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization” In Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS’20 Red Hook, NY, USA: Curran Associates Inc., 2020, pp. 7611–7623
  33. “Beyond ADMM: A Unified Client-Variance-Reduced Adaptive Federated Learning Framework” In Proceedings of the AAAI Conference on Artificial Intelligence 37.8, 2023, pp. 10175–10183 DOI: 10.1609/aaai.v37i8.26212
  34. Ziqi Wang, Yongcun Song and Enrique Zuazua “Approximate and Weighted Data Reconstruction Attack in Federated Learning” In arXiv preprint arXiv:2308.06822, 2023 DOI: 10.48550/arXiv.2308.06822
  35. Di Xiao, Jinkun Li and Min Li “Privacy-Preserving Federated Compressed Learning Against Data Reconstruction Attacks Based on Secure Data” In Neural Information Processing, Communications in Computer and Information Science Singapore: Springer Nature, 2024, pp. 325–339 DOI: 10.1007/978-981-99-8184-7˙25
  36. “ADMM without a Fixed Penalty Parameter: Faster Convergence with New Adaptive Penalization” In Advances in Neural Information Processing Systems 30 Curran Associates, Inc., 2017
  37. “Reveal Your Images: Gradient Leakage Attack against Unbiased Sampling-Based Secure Aggregation” In IEEE Transactions on Knowledge and Data Engineering, 2023, pp. 1–14 DOI: 10.1109/TKDE.2023.3271432
  38. “Implementing the Alternating Direction Method of Multipliers for Big Datasets: A Case Study of Least Absolute Shrinkage and Selection Operator” In SIAM Journal on Scientific Computing 40.5, 2018, pp. A3121–A3156 DOI: 10.1137/17M1146567
  39. “An Alternating Direction Method of Multipliers for Elliptic Equation Constrained Optimization Problem” In Science China Mathematics 60 Springer, 2017, pp. 361–378 DOI: 10.1007/s11425-015-0522-3
  40. “Federated Learning for the Internet of Things: Applications, Challenges, and Opportunities” In IEEE Internet of Things Magazine 5.1, 2022, pp. 24–29 DOI: 10.1109/IOTM.004.2100182
  41. “FedPD: A Federated Learning Framework With Adaptivity to Non-IID Data” In IEEE Transactions on Signal Processing 69, 2021, pp. 6055–6070 DOI: 10.1109/TSP.2021.3115952
  42. Shenglong Zhou and Geoffrey Ye Li “Federated Learning Via Inexact ADMM” In IEEE Transactions on Pattern Analysis and Machine Intelligence 45.8, 2023, pp. 9699–9708 DOI: 10.1109/TPAMI.2023.3243080
  43. “Parallelized Stochastic Gradient Descent” In Advances in Neural Information Processing Systems 23 Curran Associates, Inc., 2010
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Yongcun Song (12 papers)
  2. Ziqi Wang (93 papers)
  3. Enrique Zuazua (102 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com