Fast Decentralized Gradient Tracking for Federated Minimax Optimization with Local Updates (2405.04566v1)
Abstract: Federated learning (FL) for minimax optimization has emerged as a powerful paradigm for training models across distributed nodes/clients while preserving data privacy and model robustness on data heterogeneity. In this work, we delve into the decentralized implementation of federated minimax optimization by proposing \texttt{K-GT-Minimax}, a novel decentralized minimax optimization algorithm that combines local updates and gradient tracking techniques. Our analysis showcases the algorithm's communication efficiency and convergence rate for nonconvex-strongly-concave (NC-SC) minimax optimization, demonstrating a superior convergence rate compared to existing methods. \texttt{K-GT-Minimax}'s ability to handle data heterogeneity and ensure robustness underscores its significance in advancing federated learning research and applications.
- Sébastien Bubeck et al. Convex optimization: Algorithms and complexity. Foundations and Trends® in Machine Learning, 8(3-4):231–357, 2015.
- An efficient stochastic algorithm for decentralized nonconvex-strongly-concave minimax optimization. In International Conference on Artificial Intelligence and Statistics, pages 1990–1998. PMLR, 2024.
- Generative adversarial networks. Communications of the ACM, 63(11):139–144, 2020.
- The non-iid data quagmire of decentralized machine learning. In International Conference on Machine Learning, pages 4387–4398. PMLR, 2020.
- A unified theory of decentralized SGD with changing topology and local updates. In International Conference on Machine Learning, pages 5381–5393. PMLR, 2020.
- An improved analysis of gradient tracking for decentralized machine learning. Advances in Neural Information Processing Systems, 34:11422–11435, 2021.
- Advances and open problems in federated learning. Foundations and trends® in machine learning, 14(1–2):1–210, 2021.
- Communication-efficient distributed optimization in networks with gradient tracking and variance reduction. Journal of Machine Learning Research, 21(180):1–51, 2020.
- On gradient descent ascent for nonconvex-concave minimax problems. In International Conference on Machine Learning, pages 6083–6093. PMLR, 2020.
- Decentralized gradient tracking with local steps. Optimization Methods and Software, pages 1–28, 2024.
- Stochastic recursive gradient descent ascent for stochastic nonconvex-strongly-concave minimax problems. Advances in Neural Information Processing Systems, 33:20566–20577, 2020.
- PRECISION: Decentralized constrained min-max learning with low communication and sample complexities. In Proceedings of the Twenty-fourth International Symposium on Theory, Algorithmic Foundations, and Protocol Design for Mobile Networks and Mobile Computing, pages 191–200, 2023.
- On the performance of gradient tracking with local updates. In 2023 62nd IEEE Conference on Decision and Control (CDC), pages 4309–4313. IEEE, 2023.
- Decentralized saddle point problems via non-euclidean mirror prox. Optimization Methods and Software, pages 1–26, 2024.
- Federated minimax optimization with client heterogeneity. Transactions on machine learning research, 2023.
- Federated minimax optimization: Improved convergence analyses and algorithms. In International Conference on Machine Learning, pages 19683–19730. PMLR, 2022.
- A communication-efficient algorithm with linear convergence for federated minimax learning. Advances in Neural Information Processing Systems, 35:6060–6073, 2022.
- BOSE: Block-wise federated learning in heterogeneous edge computing. IEEE/ACM Transactions on Networking, 2023.
- A faster decentralized algorithm for nonconvex minimax problems. Advances in Neural Information Processing Systems, 34:25865–25877, 2021.
- Yangyang Xu. Decentralized gradient descent maximization method for composite nonconvex strongly-concave minimax problems. SIAM Journal on Optimization, 34(1):1006–1044, 2024.
- SAGDA: Achieving 𝒪(ϵ−2)𝒪superscriptitalic-ϵ2\mathcal{O}(\epsilon^{-2})caligraphic_O ( italic_ϵ start_POSTSUPERSCRIPT - 2 end_POSTSUPERSCRIPT ) communication complexity in federated min-max learning. Advances in Neural Information Processing Systems, 35:7142–7154, 2022.
- FedPAGE: Pruning adaptively toward global efficiency of heterogeneous federated learning. IEEE/ACM Transactions on Networking, 2023.
- Decentralized stochastic gradient tracking for non-convex empirical risk minimization. arXiv preprint arXiv:1909.02712, 2019.
- The complexity of nonconvex-strongly-concave minimax optimization. In Uncertainty in Artificial Intelligence, pages 482–492. PMLR, 2021.