Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fast Decentralized Gradient Tracking for Federated Minimax Optimization with Local Updates (2405.04566v1)

Published 7 May 2024 in cs.LG, cs.DC, and stat.ML

Abstract: Federated learning (FL) for minimax optimization has emerged as a powerful paradigm for training models across distributed nodes/clients while preserving data privacy and model robustness on data heterogeneity. In this work, we delve into the decentralized implementation of federated minimax optimization by proposing \texttt{K-GT-Minimax}, a novel decentralized minimax optimization algorithm that combines local updates and gradient tracking techniques. Our analysis showcases the algorithm's communication efficiency and convergence rate for nonconvex-strongly-concave (NC-SC) minimax optimization, demonstrating a superior convergence rate compared to existing methods. \texttt{K-GT-Minimax}'s ability to handle data heterogeneity and ensure robustness underscores its significance in advancing federated learning research and applications.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (24)
  1. Sébastien Bubeck et al. Convex optimization: Algorithms and complexity. Foundations and Trends® in Machine Learning, 8(3-4):231–357, 2015.
  2. An efficient stochastic algorithm for decentralized nonconvex-strongly-concave minimax optimization. In International Conference on Artificial Intelligence and Statistics, pages 1990–1998. PMLR, 2024.
  3. Generative adversarial networks. Communications of the ACM, 63(11):139–144, 2020.
  4. The non-iid data quagmire of decentralized machine learning. In International Conference on Machine Learning, pages 4387–4398. PMLR, 2020.
  5. A unified theory of decentralized SGD with changing topology and local updates. In International Conference on Machine Learning, pages 5381–5393. PMLR, 2020.
  6. An improved analysis of gradient tracking for decentralized machine learning. Advances in Neural Information Processing Systems, 34:11422–11435, 2021.
  7. Advances and open problems in federated learning. Foundations and trends® in machine learning, 14(1–2):1–210, 2021.
  8. Communication-efficient distributed optimization in networks with gradient tracking and variance reduction. Journal of Machine Learning Research, 21(180):1–51, 2020.
  9. On gradient descent ascent for nonconvex-concave minimax problems. In International Conference on Machine Learning, pages 6083–6093. PMLR, 2020.
  10. Decentralized gradient tracking with local steps. Optimization Methods and Software, pages 1–28, 2024.
  11. Stochastic recursive gradient descent ascent for stochastic nonconvex-strongly-concave minimax problems. Advances in Neural Information Processing Systems, 33:20566–20577, 2020.
  12. PRECISION: Decentralized constrained min-max learning with low communication and sample complexities. In Proceedings of the Twenty-fourth International Symposium on Theory, Algorithmic Foundations, and Protocol Design for Mobile Networks and Mobile Computing, pages 191–200, 2023.
  13. On the performance of gradient tracking with local updates. In 2023 62nd IEEE Conference on Decision and Control (CDC), pages 4309–4313. IEEE, 2023.
  14. Decentralized saddle point problems via non-euclidean mirror prox. Optimization Methods and Software, pages 1–26, 2024.
  15. Federated minimax optimization with client heterogeneity. Transactions on machine learning research, 2023.
  16. Federated minimax optimization: Improved convergence analyses and algorithms. In International Conference on Machine Learning, pages 19683–19730. PMLR, 2022.
  17. A communication-efficient algorithm with linear convergence for federated minimax learning. Advances in Neural Information Processing Systems, 35:6060–6073, 2022.
  18. BOSE: Block-wise federated learning in heterogeneous edge computing. IEEE/ACM Transactions on Networking, 2023.
  19. A faster decentralized algorithm for nonconvex minimax problems. Advances in Neural Information Processing Systems, 34:25865–25877, 2021.
  20. Yangyang Xu. Decentralized gradient descent maximization method for composite nonconvex strongly-concave minimax problems. SIAM Journal on Optimization, 34(1):1006–1044, 2024.
  21. SAGDA: Achieving 𝒪⁢(ϵ−2)𝒪superscriptitalic-ϵ2\mathcal{O}(\epsilon^{-2})caligraphic_O ( italic_ϵ start_POSTSUPERSCRIPT - 2 end_POSTSUPERSCRIPT ) communication complexity in federated min-max learning. Advances in Neural Information Processing Systems, 35:7142–7154, 2022.
  22. FedPAGE: Pruning adaptively toward global efficiency of heterogeneous federated learning. IEEE/ACM Transactions on Networking, 2023.
  23. Decentralized stochastic gradient tracking for non-convex empirical risk minimization. arXiv preprint arXiv:1909.02712, 2019.
  24. The complexity of nonconvex-strongly-concave minimax optimization. In Uncertainty in Artificial Intelligence, pages 482–492. PMLR, 2021.

Summary

We haven't generated a summary for this paper yet.