Papers
Topics
Authors
Recent
Search
2000 character limit reached

Near-optimal tensor methods for minimizing the gradient norm of convex functions and accelerated primal-dual tensor methods

Published 6 Dec 2019 in math.OC | (1912.03381v4)

Abstract: Motivated, in particular, by the entropy-regularized optimal transport problem, we consider convex optimization problems with linear equality constraints, where the dual objective has Lipschitz $p$-th order derivatives, and develop two approaches for solving such problems. The first approach is based on the minimization of the norm of the gradient in the dual problem and then the reconstruction of an approximate primal solution. Recently, Grapiglia and Nesterov in their work showed lower complexity bounds for the problem of minimizing the gradient norm of the function with Lipschitz $p$-th order derivatives. Still, the question of optimal or near-optimal methods remained open as the algorithms presented in the paper achieve suboptimal bounds only. We close this gap by proposing two near-optimal (up to logarithmic factors) methods with complexity bounds $\tilde{O}(\varepsilon{-2(p+1)/(3p+1)})$ and $\tilde{O}(\varepsilon{-2/(3p+1)})$ with respect to the initial objective residual and the distance between the starting point and solution respectively. We then apply these results (having independent interest) to our primal-dual setting. As the second approach, we propose a direct accelerated primal-dual tensor method for convex problems with linear equality constraints, where the dual objective has Lipschitz $p$-th order derivatives. For this algorithm, we prove $\tilde O (\varepsilon{-1 / (p + 1)})$ complexity in terms of the duality gap and the residual in the constraints. We illustrate the practical performance of the proposed algorithms in experiments on logistic regression, entropy-regularized optimal transport problem, and the minimal mutual information problem.

Citations (1)

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.