Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

An Improved Analysis of Gradient Tracking for Decentralized Machine Learning (2202.03836v1)

Published 8 Feb 2022 in cs.DC, cs.LG, and math.OC

Abstract: We consider decentralized machine learning over a network where the training data is distributed across $n$ agents, each of which can compute stochastic model updates on their local data. The agent's common goal is to find a model that minimizes the average of all local loss functions. While gradient tracking (GT) algorithms can overcome a key challenge, namely accounting for differences between workers' local data distributions, the known convergence rates for GT algorithms are not optimal with respect to their dependence on the mixing parameter $p$ (related to the spectral gap of the connectivity matrix). We provide a tighter analysis of the GT method in the stochastic strongly convex, convex and non-convex settings. We improve the dependency on $p$ from $\mathcal{O}(p{-2})$ to $\mathcal{O}(p{-1}c{-1})$ in the noiseless case and from $\mathcal{O}(p{-3/2})$ to $\mathcal{O}(p{-1/2}c{-1})$ in the general stochastic case, where $c \geq p$ is related to the negative eigenvalues of the connectivity matrix (and is a constant in most practical applications). This improvement was possible due to a new proof technique which could be of independent interest.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Anastasia Koloskova (18 papers)
  2. Tao Lin (168 papers)
  3. Sebastian U. Stich (66 papers)
Citations (97)

Summary

We haven't generated a summary for this paper yet.