Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Population Expansion for Training Language Models with Private Federated Learning (2307.07477v1)

Published 14 Jul 2023 in cs.LG, cs.CL, and cs.CR

Abstract: Federated learning (FL) combined with differential privacy (DP) offers ML training with distributed devices and with a formal privacy guarantee. With a large population of devices, FL with DP produces a performant model in a timely manner. However, for applications with a smaller population, not only does the model utility degrade as the DP noise is inversely proportional to population, but also the training latency increases since waiting for enough clients to become available from a smaller pool is slower. In this work, we thus propose expanding the population based on domain adaptation techniques to speed up the training and improves the final model quality when training with small populations. We empirically demonstrate that our techniques can improve the utility by 13% to 30% on real-world LLMing datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Tatsuki Koga (6 papers)
  2. Congzheng Song (23 papers)
  3. Martin Pelikan (9 papers)
  4. Mona Chitnis (5 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.