Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Distributed SGD Generalizes Well Under Asynchrony (1909.13391v1)

Published 29 Sep 2019 in stat.ML, cs.DC, and cs.LG

Abstract: The performance of fully synchronized distributed systems has faced a bottleneck due to the big data trend, under which asynchronous distributed systems are becoming a major popularity due to their powerful scalability. In this paper, we study the generalization performance of stochastic gradient descent (SGD) on a distributed asynchronous system. The system consists of multiple worker machines that compute stochastic gradients which are further sent to and aggregated on a common parameter server to update the variables, and the communication in the system suffers from possible delays. Under the algorithm stability framework, we prove that distributed asynchronous SGD generalizes well given enough data samples in the training optimization. In particular, our results suggest to reduce the learning rate as we allow more asynchrony in the distributed system. Such adaptive learning rate strategy improves the stability of the distributed algorithm and reduces the corresponding generalization error. Then, we confirm our theoretical findings via numerical experiments.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jayanth Regatti (2 papers)
  2. Gaurav Tendolkar (2 papers)
  3. Yi Zhou (438 papers)
  4. Abhishek Gupta (226 papers)
  5. Yingbin Liang (140 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.