Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Leveraging Foundation Models to Improve Lightweight Clients in Federated Learning (2311.08479v1)

Published 14 Nov 2023 in cs.LG, cs.CV, and cs.DC

Abstract: Federated Learning (FL) is a distributed training paradigm that enables clients scattered across the world to cooperatively learn a global model without divulging confidential data. However, FL faces a significant challenge in the form of heterogeneous data distributions among clients, which leads to a reduction in performance and robustness. A recent approach to mitigating the impact of heterogeneous data distributions is through the use of foundation models, which offer better performance at the cost of larger computational overheads and slower inference speeds. We introduce foundation model distillation to assist in the federated training of lightweight client models and increase their performance under heterogeneous data settings while keeping inference costs low. Our results show improvement in the global model performance on a balanced testing set, which contains rarely observed samples, even under extreme non-IID client data distributions. We conduct a thorough evaluation of our framework with different foundation model backbones on CIFAR10, with varying degrees of heterogeneous data distributions ranging from class-specific data partitions across clients to dirichlet data sampling, parameterized by values between 0.01 and 1.0.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Xidong Wu (13 papers)
  2. Wan-Yi Lin (9 papers)
  3. Devin Willmott (11 papers)
  4. Filipe Condessa (10 papers)
  5. Yufei Huang (81 papers)
  6. Zhenzhen Li (26 papers)
  7. Madan Ravi Ganesh (13 papers)
Citations (4)