Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

FedSplitX: Federated Split Learning for Computationally-Constrained Heterogeneous Clients (2310.14579v1)

Published 23 Oct 2023 in cs.LG and cs.AI

Abstract: Foundation models (FMs) have demonstrated remarkable performance in machine learning but demand extensive training data and computational resources. Federated learning (FL) addresses the challenges posed by FMs, especially related to data privacy and computational burdens. However, FL on FMs faces challenges in situations with heterogeneous clients possessing varying computing capabilities, as clients with limited capabilities may struggle to train the computationally intensive FMs. To address these challenges, we propose FedSplitX, a novel FL framework that tackles system heterogeneity. FedSplitX splits a large model into client-side and server-side components at multiple partition points to accommodate diverse client capabilities. This approach enables clients to collaborate while leveraging the server's computational power, leading to improved model performance compared to baselines that limit model size to meet the requirement of the poorest client. Furthermore, FedSplitX incorporates auxiliary networks at each partition point to reduce communication costs and delays while enhancing model performance. Our experiments demonstrate that FedSplitX effectively utilizes server capabilities to train large models, outperforming baseline approaches.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (17)
  1. On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258, 2021.
  2. Language models are few-shot learners. Advances in Neural Information Processing Systems (NeurIPS), 2020.
  3. HeteroFL: Computation and communication efficient federated learning for heterogeneous clients. In International Conference on Learning Representations (ICLR), 2021.
  4. Distributed learning of deep neural network over multiple agents. Journal of Network and Computer Applications, 116:1–8, 2018.
  5. Accelerating federated learning with split learning on locally generated losses. In International Conference on Machine Learning (ICML) 2021 Workshop on Federated Learning for User Privacy and Data Confidentiality, 2021.
  6. Deep residual learning for image recognition. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
  7. FjORD: Fair and accurate federated learning under heterogeneous targets with ordered dropout. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
  8. NeFL: Nested federated learning for heterogeneous clients, 2023.
  9. DepthFL : Depthwise federated learning for heterogeneous clients. In International Conference on Learning Representations (ICLR), 2023.
  10. CIFAR-10 (Canadian Institute for Advanced Research). http://www.cs.toronto.edu/~kriz/cifar.html.
  11. Communication-efficient learning of deep networks from decentralized data. In International Conference on Artificial Intelligence and Statistics (AISTATS), 2017.
  12. Sebastian Ruder. An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747, 2016.
  13. Splitfed: When federated learning meets split learning. AAAI Conference on Artificial Intelligence (AAAI), 2022.
  14. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
  15. Split learning for health: Distributed deep learning without sharing raw patient data. arXiv preprint arXiv:1812.00564, 2018.
  16. Will we run out of data? an analysis of the limits of scaling datasets in machine learning. arXiv preprint arXiv:2211.04325, 2022.
  17. When foundation model meets federated learning: Motivations, challenges, and future directions. arXiv preprint arXiv:2306.15546, 2023.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Jiyun Shin (1 paper)
  2. Jinhyun Ahn (1 paper)
  3. Honggu Kang (4 papers)
  4. Joonhyuk Kang (59 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.