Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Optimizing the Communication-Accuracy Trade-off in Federated Learning with Rate-Distortion Theory (2201.02664v3)

Published 7 Jan 2022 in cs.LG, cs.DC, cs.IT, math.IT, and stat.ML

Abstract: A significant bottleneck in federated learning (FL) is the network communication cost of sending model updates from client devices to the central server. We present a comprehensive empirical study of the statistics of model updates in FL, as well as the role and benefits of various compression techniques. Motivated by these observations, we propose a novel method to reduce the average communication cost, which is near-optimal in many use cases, and outperforms Top-K, DRIVE, 3LC and QSGD on Stack Overflow next-word prediction, a realistic and challenging FL benchmark. This is achieved by examining the problem using rate-distortion theory, and proposing distortion as a reliable proxy for model accuracy. Distortion can be more effectively used for optimizing the trade-off between model performance and communication cost across clients. We demonstrate empirically that in spite of the non-i.i.d. nature of federated learning, the rate-distortion frontier is consistent across datasets, optimizers, clients and training rounds.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Nicole Mitchell (7 papers)
  2. Johannes Ballé (29 papers)
  3. Zachary Charles (33 papers)
  4. Jakub Konečný (28 papers)
Citations (19)

Summary

We haven't generated a summary for this paper yet.