Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Memory-Efficient Split Federated Learning for LLM Fine-Tuning on Heterogeneous Mobile Devices (2506.02940v1)

Published 3 Jun 2025 in cs.DC

Abstract: In this paper, we propose an edge-assisted split federated learning framework to facilitate LLM fine-tuning on heterogeneous mobile devices while alleviating memory pressures on both mobile devices and the edge server. Specifically, mobile devices perform low-rank adaptation (LoRA) fine-tuning on only a subset of lower layers of the pre-trained LLM, tailored to their individual capacities. On the server, a full LLM is maintained, and the corresponding LoRA modules are selectively fine-tuned in a sequential manner for each device. To further enhance training efficiency, we propose a server-side training scheduling method that optimizes the processing order of devices for accelerating fine-tuning. Extensive experiments demonstrate that compared to the baselines, our scheme can reduce 79\% memory footprint and 6\% training time while achieving comparable performance.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Xiaopei Chen (4 papers)
  2. Liang Li (297 papers)
  3. Fei Ji (17 papers)
  4. Wen Wu (103 papers)

Summary

We haven't generated a summary for this paper yet.