FedBaF: Federated Learning Aggregation Biased by a Foundation Model (2410.18352v3)
Abstract: Foundation models are now a major focus of leading technology organizations due to their ability to generalize across diverse tasks. Existing approaches for adapting foundation models to new applications often rely on Federated Learning (FL) and disclose the foundation model weights to clients when using it to initialize the global model. While these methods ensure client data privacy, they compromise model and information security. In this paper, we introduce Federated Learning Aggregation Biased by a Foundation Model (FedBaF), a novel method for dynamically integrating pre-trained foundation model weights during the FL aggregation phase. Unlike conventional methods, FedBaF preserves the confidentiality of the foundation model while still leveraging its power to train more accurate models, especially in non-IID and adversarial scenarios. Our comprehensive experiments use Pre-ResNet and foundation models like Vision Transformer to demonstrate that FedBaF not only matches, but often surpasses the test accuracy of traditional weight initialization methods by up to 11.4% in IID and up to 15.8% in non-IID settings. Additionally, FedBaF applied to a Transformer-based LLM significantly reduced perplexity by up to 39.2%.
- W. Zhuang, C. Chen, and L. Lyu, “When foundation model meets federated learning: Motivations, challenges, and future directions,” 2024.
- R. Bommasani, D. A. Hudson, E. Adeli, R. Altman, S. Arora, S. von Arx, M. S. Bernstein, J. Bohg, A. Bosselut, E. Brunskill, et al., “On the opportunities and risks of foundation models,” arXiv preprint arXiv:2108.07258, 2021.
- H.-Y. Chen, C.-H. Tu, Z. Li, H. W. Shen, and W.-L. Chao, “On the importance and applicability of pre-training for federated learning,” in The Eleventh International Conference on Learning Representations, 2022.
- X. Han, Z. Zhang, N. Ding, Y. Gu, X. Liu, Y. Huo, J. Qiu, Y. Yao, A. Zhang, L. Zhang, et al., “Pre-trained models: Past, present and future,” AI Open, vol. 2, pp. 225–250, 2021.
- M. Duan, D. Liu, X. Ji, Y. Wu, L. Liang, X. Chen, Y. Tan, and A. Ren, “Flexible clustered federated learning for client-level data distribution shift,” IEEE Transactions on Parallel and Distributed Systems, vol. 33, no. 11, pp. 2661–2674, 2021.
- M. Joshi, A. Pal, and M. Sankarasubbu, “Federated learning for healthcare domain - pipeline, applications and challenges,” ACM Trans. Comput. Healthcare, vol. 3, nov 2022.
- J. Yosinski, J. Clune, Y. Bengio, and H. Lipson, “How transferable are features in deep neural networks?,” Advances in Neural Information Processing Systems, vol. 27, 2014.
- S. Pranav and J. M. Moura, “Peer-to-peer deep learning for beyond-5g iot,” arXiv preprint arXiv:2310.18861, 2023.
- Q. Li, Z. Wen, Z. Wu, S. Hu, N. Wang, Y. Li, X. Liu, and B. He, “A survey on federated learning systems: Vision, hype and reality for data privacy and protection,” IEEE Transactions on Knowledge and Data Engineering, 2021.
- B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, “Communication-efficient learning of deep networks from decentralized data,” in Artificial Intelligence and Statistics, pp. 1273–1282, PMLR, 2017.
- Cham: Springer International Publishing, 2020.
- D. C. Nguyen, M. Ding, P. N. Pathirana, A. Seneviratne, J. Li, and H. V. Poor, “Federated learning for internet of things: A comprehensive survey,” IEEE Communications Surveys & Tutorials, vol. 23, no. 3, pp. 1622–1658, 2021.
- M. Siew, H. Zhang, J.-I. Park, Y. Liu, Y. Ruan, L. Su, S. Ioannidis, E. Yeh, and C. Joe-Wong, “Fair concurrent training of multiple models in federated learning,” arXiv preprint arXiv:2404.13841, 2024.
- A. Singh, P. Vepakomma, O. Gupta, and R. Raskar, “‘detailed comparison of communication efficiency of split learning and federated learning,” arXiv preprint arXiv:1909.09145, 2019.
- J. Nguyen, K. Malik, M. Sanjabi, and M. Rabbat, “Where to begin? exploring the impact of pre-training and initialization in federated learning,” arXiv preprint arXiv:2206.15387, 2022.
- S. Dayal, D. Alhadidi, A. Abbasi Tadi, and N. Mohammed, “Comparative analysis of membership inference attacks in federated learning,” in Proceedings of the 27th International Database Engineered Applications Symposium, pp. 185–192, 2023.
- H. Hu, Z. Salcic, L. Sun, G. Dobbie, P. S. Yu, and X. Zhang, “Membership inference attacks on machine learning: A survey,” ACM Computing Surveys (CSUR), vol. 54, no. 11s, pp. 1–37, 2022.
- X. Wang, N. Wang, L. Wu, Z. Guan, X. Du, and M. Guizani, “Gbmia: Gradient-based membership inference attack in federated learning,” in ICC 2023-IEEE International Conference on Communications, pp. 5066–5071, IEEE, 2023.
- M. Fredrikson, S. Jha, and T. Ristenpart, “Model inversion attacks that exploit confidence information and basic countermeasures,” in Proceedings of the 22nd ACM SIGSAC conference on computer and communications security, pp. 1322–1333, 2015.
- J. Li, A. S. Rakin, X. Chen, Z. He, D. Fan, and C. Chakrabarti, “Ressfl: A resistance transfer framework for defending model inversion attack in split federated learning,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10194–10202, 2022.
- Y. Zhang, R. Jia, H. Pei, W. Wang, B. Li, and D. Song, “The secret revealer: Generative model-inversion attacks against deep neural networks,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 253–261, 2020.
- E. Bagdasaryan, A. Veit, Y. Hua, D. Estrin, and V. Shmatikov, “How to backdoor federated learning,” in International conference on artificial intelligence and statistics, pp. 2938–2948, PMLR, 2020.
- T. Kim, J. Li, N. Madaan, S. Singh, and C. Joe-Wong, “Adversarial robustness unhardening via backdoor attacks in federated learning,” in NeurIPS 2023 Workshop on Backdoors in Deep Learning-The Good, the Bad, and the Ugly, 2023.
- J. Yu, Y. Wang, C. Zhao, B. Ghanem, and J. Zhang, “Freedom: Training-free energy-guided conditional diffusion model,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 23174–23184, October 2023.
- S. Pranav and J. M. Moura, “Peer-to-peer learning+ consensus with non-iid data,” in 2023 57th Asilomar Conference on Signals, Systems, and Computers, pp. 709–713, IEEE, 2023.
- Y. Zhao, M. Li, L. Lai, N. Suda, D. Civin, and V. Chandra, “Federated learning with non-iid data,” arXiv preprint arXiv:1806.00582, 2018.
- Z. Li, K. Ren, X. Jiang, B. Li, H. Zhang, and D. Li, “Domain generalization using pretrained models without fine-tuning,” arXiv preprint arXiv:2203.04600, 2022.
- Y. Tan, G. Long, J. Ma, L. Liu, T. Zhou, and J. Jiang, “Federated learning from pre-trained models: A contrastive learning approach,” Advances in Neural Information Processing Systems, vol. 35, pp. 19332–19344, 2022.
- P. Xu, X. Zhu, and D. A. Clifton, “Multimodal learning with transformers: A survey,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 10, pp. 12113–12132, 2023.
- T. Kenneweg, P. Kenneweg, and B. Hammer, “Foundation model vision transformers are great tracking backbones,” in 2024 International Conference on Artificial Intelligence, Computer, Data Sciences and Applications (ACDSA), pp. 1–6, IEEE, 2024.
- T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith, “Federated optimization in heterogeneous networks,” Proceedings of Machine learning and systems, vol. 2, pp. 429–450, 2020.
- Springer, 2021.
- S. P. Karimireddy, S. Kale, M. Mohri, S. Reddi, S. Stich, and A. T. Suresh, “Scaffold: Stochastic controlled averaging for federated learning,” in International conference on machine learning, pp. 5132–5143, PMLR, 2020.
- R. Vaz, “Rome weather classification.” https://www.kaggle.com/datasets/rogeriovaz/rome-weather-classification, 2021. Accessed: 2024-05-21.
- A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
- C. Zhou, Q. Li, C. Li, J. Yu, Y. Liu, G. Wang, K. Zhang, C. Ji, Q. Yan, L. He, et al., “A comprehensive survey on pretrained foundation models: A history from bert to chatgpt,” arXiv preprint arXiv:2302.09419, 2023.
- H. Xiao, “Weather phenomenon database (WEAPD),” 2021.
- J.-I. Park and C. Joe-Wong, “Federated learning with flexible architectures,” in Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 143–161, Springer, 2024.
- E. Diao, J. Ding, and V. Tarokh, “Heterofl: Computation and communication efficient federated learning for heterogeneous clients,” arXiv preprint arXiv:2010.01264, 2020.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.