Papers
Topics
Authors
Recent
2000 character limit reached

A General Convergence Result for the Exponentiated Gradient Method (1705.09628v1)

Published 26 May 2017 in math.OC

Abstract: The batch exponentiated gradient (EG) method provides a principled approach to convex smooth minimization on the probability simplex or the space of quantum density matrices. However, it is not always guaranteed to converge. Existing convergence analyses of the EG method require certain quantitative smoothness conditions on the loss function, e.g., Lipschitz continuity of the loss function or its gradient, but those conditions may not hold in important applications. In this paper, we prove that the EG method with Armijo line search always converges for any convex loss function with a locally Lipschitz continuous gradient. Because of our convergence guarantee, the EG method with Armijo line search becomes the fastest guaranteed-to-converge algorithm for maximum-likelihood quantum state estimation, on the real datasets we have.

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Paper to Video (Beta)

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.