Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adaptive Differential Filters for Fast and Communication-Efficient Federated Learning (2204.04424v1)

Published 9 Apr 2022 in cs.LG, cs.AI, cs.CV, and cs.DC

Abstract: Federated learning (FL) scenarios inherently generate a large communication overhead by frequently transmitting neural network updates between clients and server. To minimize the communication cost, introducing sparsity in conjunction with differential updates is a commonly used technique. However, sparse model updates can slow down convergence speed or unintentionally skip certain update aspects, e.g., learned features, if error accumulation is not properly addressed. In this work, we propose a new scaling method operating at the granularity of convolutional filters which 1) compensates for highly sparse updates in FL processes, 2) adapts the local models to new data domains by enhancing some features in the filter space while diminishing others and 3) motivates extra sparsity in updates and thus achieves higher compression ratios, i.e., savings in the overall data transfer. Compared to unscaled updates and previous work, experimental results on different computer vision tasks (Pascal VOC, CIFAR10, Chest X-Ray) and neural networks (ResNets, MobileNets, VGGs) in uni-, bidirectional and partial update FL settings show that the proposed method improves the performance of the central server model while converging faster and reducing the total amount of transmitted data by up to 377 times.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Daniel Becking (5 papers)
  2. Heiner Kirchhoffer (2 papers)
  3. Gerhard Tech (1 paper)
  4. Paul Haase (3 papers)
  5. Karsten Müller (12 papers)
  6. Heiko Schwarz (4 papers)
  7. Wojciech Samek (144 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.