Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fundamental Limits of Approximate Gradient Coding (1901.08166v1)

Published 23 Jan 2019 in cs.DC

Abstract: It has been established that when the gradient coding problem is distributed among $n$ servers, the computation load (number of stored data partitions) of each worker is at least $s+1$ in order to resists $s$ stragglers. This scheme incurs a large overhead when the number of stragglers $s$ is large. In this paper, we focus on a new framework called \emph{approximate gradient coding} to mitigate stragglers in distributed learning. We show that, to exactly recover the gradient with high probability, the computation load is lower bounded by $O(\log(n)/\log(n/s))$. We also propose a code that exactly matches such lower bound. We identify a fundamental three-fold tradeoff for any approximate gradient coding scheme $d\geq O(\log(1/\epsilon)/\log(n/s))$, where $d$ is the computation load, $\epsilon$ is the error of gradient. We give an explicit code construction based on random edge removal process that achieves the derived tradeoff. We implement our schemes and demonstrate the advantage of the approaches over the current fastest gradient coding strategies.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Sinong Wang (45 papers)
  2. Jiashang Liu (3 papers)
  3. Ness Shroff (51 papers)
Citations (42)

Summary

We haven't generated a summary for this paper yet.