Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Escape saddle points faster on manifolds via perturbed Riemannian stochastic recursive gradient (2010.12191v2)

Published 23 Oct 2020 in math.OC and cs.LG

Abstract: In this paper, we propose a variant of Riemannian stochastic recursive gradient method that can achieve second-order convergence guarantee and escape saddle points using simple perturbation. The idea is to perturb the iterates when gradient is small and carry out stochastic recursive gradient updates over tangent space. This avoids the complication of exploiting Riemannian geometry. We show that under finite-sum setting, our algorithm requires $\widetilde{\mathcal{O}}\big( \frac{ \sqrt{n}}{\epsilon2} + \frac{\sqrt{n} }{\delta4} + \frac{n}{\delta3}\big)$ stochastic gradient queries to find a $(\epsilon, \delta)$-second-order critical point. This strictly improves the complexity of perturbed Riemannian gradient descent and is superior to perturbed Riemannian accelerated gradient descent under large-sample settings. We also provide a complexity of $\widetilde{\mathcal{O}} \big( \frac{1}{\epsilon3} + \frac{1}{\delta3 \epsilon2} + \frac{1}{\delta4 \epsilon} \big)$ for online optimization, which is novel on Riemannian manifold in terms of second-order convergence using only first-order information.

Citations (4)

Summary

We haven't generated a summary for this paper yet.