Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

When Federated Learning Meets Pre-trained Language Models' Parameter-Efficient Tuning Methods (2212.10025v2)

Published 20 Dec 2022 in cs.LG and cs.CL

Abstract: With increasing privacy concerns on data, recent studies have made significant progress using federated learning (FL) on privacy-sensitive NLP tasks. Much literature suggests fully fine-tuning pre-trained LLMs (PLMs) in the FL paradigm can mitigate the data heterogeneity problem and close the performance gap with centralized training. However, large PLMs bring the curse of prohibitive communication overhead and local model adaptation costs for the FL system. To this end, we introduce various parameter-efficient tuning (PETuning) methods into federated learning. Specifically, we provide a holistic empirical study of representative PLMs tuning methods in FL. The experimental results cover the analysis of data heterogeneity levels, data scales, and different FL scenarios. Overall communication overhead can be significantly reduced by locally tuning and globally aggregating lightweight model parameters while maintaining acceptable performance in various FL settings. To facilitate the research of PETuning in FL, we also develop a federated tuning framework FedPETuning, which allows practitioners to exploit different PETuning methods under the FL training paradigm conveniently. The source code is available at \url{https://github.com/iezhuozhuo/FedETuning/tree/deltaTuning}.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Zhuo Zhang (42 papers)
  2. Yuanhang Yang (8 papers)
  3. Yong Dai (33 papers)
  4. Lizhen Qu (68 papers)
  5. Zenglin Xu (145 papers)
Citations (45)
Github Logo Streamline Icon: https://streamlinehq.com

GitHub