2000 character limit reached
Minimax Efficient Finite-Difference Stochastic Gradient Estimators Using Black-Box Function Evaluations
Published 8 Jul 2020 in math.ST, stat.ME, and stat.TH | (2007.04443v2)
Abstract: Standard approaches to stochastic gradient estimation, with only noisy black-box function evaluations, use the finite-difference method or its variants. While natural, it is open to our knowledge whether their statistical accuracy is the best possible. This paper argues so by showing that central finite-difference is a nearly minimax optimal zeroth-order gradient estimator for a suitable class of objective functions and mean squared risk, among both the class of linear estimators and the much larger class of all (nonlinear) estimators.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.