Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Linearly Convergent Algorithm with Variance Reduction for Distributed Stochastic Optimization (2002.03269v2)

Published 9 Feb 2020 in math.OC and cs.MA

Abstract: This paper considers a distributed stochastic strongly convex optimization, where agents connected over a network aim to cooperatively minimize the average of all agents' local cost functions. Due to the stochasticity of gradient estimation and distributedness of local objective, fast linearly convergent distributed algorithms have not been achieved yet. This work proposes a novel distributed stochastic gradient tracking algorithm with variance reduction, where the local gradients are estimated by an increasing batch-size of sampled gradients. With an undirected connected communication graph and a geometrically increasing batch-size, the iterates are shown to converge in mean to the optimal solution at a geometric rate (achieving linear convergence). The iteration, communication, and oracle complexity for obtaining an $\epsilon$-optimal solution are established as well. Particulary, the communication complexity is $\mathcal{O}(\ln (1/\epsilon))$ while the oracle complexity (number of sampled gradients) is $\mathcal{O}(1/\epsilon2)$, which is of the same order as that of centralized approaches. Hence, the proposed scheme is communication-efficient without requiring extra sampled gradients. Numerical simulations are given to demonstrate the theoretic results.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Jinlong Lei (31 papers)
  2. Peng Yi (52 papers)
  3. Jie Chen (602 papers)
  4. Yiguang Hong (87 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.