Papers
Topics
Authors
Recent
Search
2000 character limit reached

Optimal score estimation via empirical Bayes smoothing

Published 12 Feb 2024 in math.ST, stat.ML, and stat.TH | (2402.07747v2)

Abstract: We study the problem of estimating the score function of an unknown probability distribution $\rho*$ from $n$ independent and identically distributed observations in $d$ dimensions. Assuming that $\rho*$ is subgaussian and has a Lipschitz-continuous score function $s*$, we establish the optimal rate of $\tilde \Theta(n{-\frac{2}{d+4}})$ for this estimation problem under the loss function $|\hat s - s|2_{L2(\rho^)}$ that is commonly used in the score matching literature, highlighting the curse of dimensionality where sample complexity for accurate score estimation grows exponentially with the dimension $d$. Leveraging key insights in empirical Bayes theory as well as a new convergence rate of smoothed empirical distribution in Hellinger distance, we show that a regularized score estimator based on a Gaussian kernel attains this rate, shown optimal by a matching minimax lower bound. We also discuss extensions to estimating $\beta$-H\"older continuous scores with $\beta \leq 1$, as well as the implication of our theory on the sample complexity of score-based generative models.

Citations (14)

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 1 like about this paper.