2000 character limit reached
A Stochastic Gradient Method with an Exponential Convergence Rate for Finite Training Sets (1202.6258v4)
Published 28 Feb 2012 in math.OC and cs.LG
Abstract: We propose a new stochastic gradient method for optimizing the sum of a finite set of smooth functions, where the sum is strongly convex. While standard stochastic gradient methods converge at sublinear rates for this problem, the proposed method incorporates a memory of previous gradient values in order to achieve a linear convergence rate. In a machine learning context, numerical experiments indicate that the new algorithm can dramatically outperform standard algorithms, both in terms of optimizing the training error and reducing the test error quickly.
- Nicolas Le Roux (41 papers)
- Mark Schmidt (74 papers)
- Francis Bach (249 papers)