Faster convergence of momentum-based methods in nonconvex optimization
Determine whether momentum-based methods, such as Nesterov's accelerated gradient descent, achieve faster convergence rates than standard gradient descent for nonconvex optimization problems, specifically with respect to finding second-order stationary points that avoid strict saddle points.
References
It is open as to whether momentum-based methods yield faster rates in the nonconvex setting, specifically when we consider the convergence criterion of second-order stationarity.
— Accelerated Gradient Descent Escapes Saddle Points Faster than Gradient Descent
(1711.10456 - Jin et al., 2017) in Section 1 (Introduction)