Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GPT-FL: Generative Pre-trained Model-Assisted Federated Learning (2306.02210v4)

Published 3 Jun 2023 in cs.LG and cs.DC

Abstract: In this work, we propose GPT-FL, a generative pre-trained model-assisted federated learning (FL) framework. At its core, GPT-FL leverages generative pre-trained models to generate diversified synthetic data. These generated data are used to train a downstream model on the server, which is then fine-tuned with private client data under the standard FL framework. We show that GPT-FL consistently outperforms state-of-the-art FL methods in terms of model test accuracy, communication efficiency, and client sampling efficiency. Through comprehensive ablation analysis across various data modalities, we discover that the downstream model generated by synthetic data plays a crucial role in controlling the direction of gradient diversity during FL training, which enhances convergence speed and contributes to the notable accuracy boost observed with GPT-FL. Also, regardless of whether the target data falls within or outside the domain of the pre-trained generative model, GPT-FL consistently achieves significant performance gains, surpassing the results obtained by models trained solely with FL or synthetic data. The code is available at https://github.com/AvestimehrResearchGroup/GPT-FL.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (42)
  1. Fedrolex: Model-heterogeneous federated learning with rolling sub-model extraction. ArXiv, abs/2212.01548, 2022.
  2. Speecht5: Unified-modal encoder-decoder pre-training for spoken language processing. In Annual Meeting of the Association for Computational Linguistics, 2021.
  3. Practical secure aggregation for federated learning on user-held data. In NIPS Workshop on Private Multi-Party Machine Learning, 2016. URL https://arxiv.org/abs/1611.04482.
  4. Practical secure aggregation for privacy-preserving machine learning. Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, 2017.
  5. Towards federated learning at scale: System design. ArXiv, abs/1902.01046, 2019.
  6. Heterogeneous ensemble knowledge transfer for training large models in federated learning. In International Joint Conference on Artificial Intelligence, 2022.
  7. A method to reveal speaker identity in distributed asr training, and how to counter it. ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.  4338–4342, 2021.
  8. Federated heavy hitter recovery under linear sketching. ArXiv, abs/2307.13347, 2023. URL https://api.semanticscholar.org/CorpusID:260154975.
  9. Invertible bloom lookup tables. 2011 49th Annual Allerton Conference on Communication, Control, and Computing (Allerton), pp.  792–799, 2011. URL https://api.semanticscholar.org/CorpusID:11589877.
  10. Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.  770–778, 2015.
  11. Is synthetic data from generative models ready for image recognition? ArXiv, abs/2210.07574, 2022.
  12. Evaluating gradient inversion attacks and defenses in federated learning. In Neural Information Processing Systems, 2021.
  13. Distillation-based semi-supervised federated learning for communication-efficient collaborative training with non-iid private data. IEEE Transactions on Mobile Computing, 22:191–205, 2020.
  14. Scaffold: Stochastic controlled averaging for federated learning. In International Conference on Machine Learning, 2019.
  15. Dataset condensation via efficient synthetic-data parameterization. In International Conference on Machine Learning, 2022.
  16. Alex Krizhevsky. Learning multiple layers of features from tiny images. 2009. URL https://api.semanticscholar.org/CorpusID:18268744.
  17. Training neural speech recognition systems with synthetic speech augmentation. arXiv preprint arXiv:1811.00707, 2018.
  18. Model-contrastive federated learning. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  10708–10717, 2021.
  19. Ensemble distillation for robust model fusion in federated learning. ArXiv, abs/2006.07242, 2020.
  20. Audioldm: Text-to-audio generation with latent diffusion models. ArXiv, abs/2301.12503, 2023.
  21. Decoupled weight decay regularization. In International Conference on Learning Representations, 2017.
  22. Communication-efficient learning of deep networks from decentralized data. In International Conference on Artificial Intelligence and Statistics, 2016.
  23. Where to begin? on the impact of pre-training and initialization in federated learning. ArXiv, abs/2210.08090, 2022.
  24. Automated flower classification over a large number of classes. 2008 Sixth Indian Conference on Computer Vision, Graphics & Image Processing, pp.  722–729, 2008.
  25. Dynafed: Tackling client data heterogeneity with global dynamics. arXiv preprint arXiv:2211.10878, 2022.
  26. Karol J. Piczak. ESC: Dataset for Environmental Sound Classification. In Proceedings of the 23rd Annual ACM Conference on Multimedia, pp.  1015–1018. ACM Press. ISBN 978-1-4503-3459-4. doi: 10.1145/2733373.2806390. URL http://dl.acm.org/citation.cfm?doid=2733373.2806390.
  27. Adaptive federated optimization. ArXiv, abs/2003.00295, 2020.
  28. High-resolution image synthesis with latent diffusion models. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  10674–10685, 2021.
  29. Federated optimization in heterogeneous networks. arXiv: Learning, 2018.
  30. Laion-5b: An open large-scale dataset for training next generation image-text models. ArXiv, abs/2210.08402, 2022.
  31. Diversity is definitely needed: Improving model-agnostic zero-shot classification via stable diffusion. 2023.
  32. Very deep convolutional networks for large-scale image recognition. CoRR, abs/1409.1556, 2014.
  33. Lightsecagg: a lightweight and versatile design for secure aggregation in federated learning. In Conference on Machine Learning and Systems, 2021.
  34. Does knowledge distillation really work? ArXiv, abs/2106.05945, 2021.
  35. TensorFlow. Private heavy hitters, 2023. URL https://www.tensorflow.org/federated/tutorials/private_heavy_hitters.
  36. Attack of the tails: Yes, you really can backdoor federated learning. ArXiv, abs/2007.05084, 2020.
  37. Pete Warden. Speech commands: A dataset for limited-vocabulary speech recognition. ArXiv, abs/1804.03209, 2018.
  38. Gradient diversity: a key ingredient for scalable distributed learning. In International Conference on Artificial Intelligence and Statistics, pp.  1998–2007. PMLR, 2018.
  39. Fine-tuning global model via data-free knowledge distillation for non-iid federated learning. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  10164–10173, 2022.
  40. Federated learning for the internet of things: Applications, challenges, and opportunities. IEEE Internet of Things Magazine, 5:24–29, 2021.
  41. Fedaudio: A federated learning benchmark for audio tasks. In ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.  1–5, 2023. doi: 10.1109/ICASSP49357.2023.10096500.
  42. Data-free knowledge distillation for heterogeneous federated learning. Proceedings of machine learning research, 139:12878–12889, 2021.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Tuo Zhang (46 papers)
  2. Tiantian Feng (61 papers)
  3. Samiul Alam (15 papers)
  4. Dimitrios Dimitriadis (32 papers)
  5. Mi Zhang (85 papers)
  6. Shrikanth S. Narayanan (14 papers)
  7. Salman Avestimehr (116 papers)
  8. Sunwoo Lee (32 papers)
Citations (19)

Summary

We haven't generated a summary for this paper yet.