Papers
Topics
Authors
Recent
Search
2000 character limit reached

Convergence properties of an Objective-Function-Free Optimization regularization algorithm, including an $\mathcal{O}(ε^{-3/2})$ complexity bound

Published 18 Mar 2022 in math.OC | (2203.09947v2)

Abstract: An adaptive regularization algorithm for unconstrained nonconvex optimization is presented in which the objective function is never evaluated, but only derivatives are used. This algorithm belongs to the class of adaptive regularization methods, for which optimal worst-case complexity results are known for the standard framework where the objective function is evaluated. It is shown in this paper that these excellent complexity bounds are also valid for the new algorithm, despite the fact that significantly less information is used. In particular, it is shown that, if derivatives of degree one to $p$ are used, the algorithm will find a $\epsilon_1$-approximate first-order minimizer in at most $O(\epsilon_1{-(p+1)/p})$ iterations, and an $(\epsilon_1,\epsilon_2)$-approximate second-order minimizer in at most $O(\max[\epsilon{-(p+1)/p},\epsilon_2{-(p+1)/(p-1)}])$ iterations. As a special case, the new algorithm using first and second derivatives, when applied to functions with Lipschitz continuous Hessian, will find an iterate $x_k$ at which the gradient's norm is less than $\epsilon_1$ in at most $O(\epsilon_1{-3/2})$ iterations.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (3)

Collections

Sign up for free to add this paper to one or more collections.