Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Optimal score estimation via empirical Bayes smoothing (2402.07747v2)

Published 12 Feb 2024 in math.ST, stat.ML, and stat.TH

Abstract: We study the problem of estimating the score function of an unknown probability distribution $\rho*$ from $n$ independent and identically distributed observations in $d$ dimensions. Assuming that $\rho*$ is subgaussian and has a Lipschitz-continuous score function $s*$, we establish the optimal rate of $\tilde \Theta(n{-\frac{2}{d+4}})$ for this estimation problem under the loss function $|\hat s - s|2_{L2(\rho^)}$ that is commonly used in the score matching literature, highlighting the curse of dimensionality where sample complexity for accurate score estimation grows exponentially with the dimension $d$. Leveraging key insights in empirical Bayes theory as well as a new convergence rate of smoothed empirical distribution in Hellinger distance, we show that a regularized score estimator based on a Gaussian kernel attains this rate, shown optimal by a matching minimax lower bound. We also discuss extensions to estimating $\beta$-H\"older continuous scores with $\beta \leq 1$, as well as the implication of our theory on the sample complexity of score-based generative models.

Citations (14)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets