Papers
Topics
Authors
Recent
Search
2000 character limit reached

A tight bound on the stepsize of the decentralized gradient descent

Published 10 Mar 2023 in math.OC, cs.SY, and eess.SY | (2303.05755v1)

Abstract: In this paper, we consider the decentralized gradinet descent (DGD) given by \begin{equation*} x_i (t+1) = \sum_{j=1}m w_{ij} x_j (t) - \alpha (t) \nabla f_i (x_i (t)). \end{equation*} We find a sharp range of the stepsize $\alpha (t)>0$ such that the sequence ${x_i (t)}$ is uniformly bounded when the aggregate cost $f$ is assumed be strongly convex with smooth local costs which might be non-convex. Precisely, we find a tight bound $\alpha_0 >0$ such that the states of the DGD algorithm is uniformly bounded for non-increasing sequence $\alpha (t)$ satisfying $\alpha (0) \leq \alpha_0$. The theoretical results are also verified by numerical experiments.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (1)

Collections

Sign up for free to add this paper to one or more collections.