Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

$\texttt{DeepSqueeze}$: Decentralization Meets Error-Compensated Compression (1907.07346v2)

Published 17 Jul 2019 in cs.DC, cs.LG, and stat.ML

Abstract: Communication is a key bottleneck in distributed training. Recently, an \emph{error-compensated} compression technology was particularly designed for the \emph{centralized} learning and receives huge successes, by showing significant advantages over state-of-the-art compression based methods in saving the communication cost. Since the \emph{decentralized} training has been witnessed to be superior to the traditional \emph{centralized} training in the communication restricted scenario, therefore a natural question to ask is "how to apply the error-compensated technology to the decentralized learning to further reduce the communication cost." However, a trivial extension of compression based centralized training algorithms does not exist for the decentralized scenario. key difference between centralized and decentralized training makes this extension extremely non-trivial. In this paper, we propose an elegant algorithmic design to employ error-compensated stochastic gradient descent for the decentralized scenario, named $\texttt{DeepSqueeze}$. Both the theoretical analysis and the empirical study are provided to show the proposed $\texttt{DeepSqueeze}$ algorithm outperforms the existing compression based decentralized learning algorithms. To the best of our knowledge, this is the first time to apply the error-compensated compression to the decentralized learning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Hanlin Tang (34 papers)
  2. Xiangru Lian (18 papers)
  3. Shuang Qiu (46 papers)
  4. Lei Yuan (34 papers)
  5. Ce Zhang (215 papers)
  6. Tong Zhang (569 papers)
  7. Ji Liu (285 papers)
Citations (48)

Summary

We haven't generated a summary for this paper yet.