Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Projection-free Distributed Online Learning with Sublinear Communication Complexity (2103.11102v2)

Published 20 Mar 2021 in cs.LG and stat.ML

Abstract: To deal with complicated constraints via locally light computations in distributed online learning, a recent study has presented a projection-free algorithm called distributed online conditional gradient (D-OCG), and achieved an $O(T{3/4})$ regret bound for convex losses, where $T$ is the number of total rounds. However, it requires $T$ communication rounds, and cannot utilize the strong convexity of losses. In this paper, we propose an improved variant of D-OCG, namely D-BOCG, which can attain the same $O(T{3/4})$ regret bound with only $O(\sqrt{T})$ communication rounds for convex losses, and a better regret bound of $O(T{2/3}(\log T){1/3})$ with fewer $O(T{1/3}(\log T){2/3})$ communication rounds for strongly convex losses. The key idea is to adopt a delayed update mechanism that reduces the communication complexity, and redefine the surrogate loss function in D-OCG for exploiting the strong convexity. Furthermore, we provide lower bounds to demonstrate that the $O(\sqrt{T})$ communication rounds required by D-BOCG are optimal (in terms of $T$) for achieving the $O(T{3/4})$ regret with convex losses, and the $O(T{1/3}(\log T){2/3})$ communication rounds required by D-BOCG are near-optimal (in terms of $T$) for achieving the $O(T{2/3}(\log T){1/3})$ regret with strongly convex losses up to polylogarithmic factors. Finally, to handle the more challenging bandit setting, in which only the loss value is available, we incorporate the classical one-point gradient estimator into D-BOCG, and obtain similar theoretical guarantees.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Yuanyu Wan (23 papers)
  2. Guanghui Wang (179 papers)
  3. Wei-Wei Tu (29 papers)
  4. Lijun Zhang (239 papers)
Citations (12)

Summary

We haven't generated a summary for this paper yet.