Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
184 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

The fast rate of convergence of the smooth adapted Wasserstein distance (2503.10827v1)

Published 13 Mar 2025 in math.PR

Abstract: Estimating a $d$-dimensional distribution $\mu$ by the empirical measure $\hat{\mu}_n$ of its samples is an important task in probability theory, statistics and machine learning. It is well known that $\mathbb{E}[\mathcal{W}_p(\hat{\mu}_n, \mu)]\lesssim n{-1/d}$ for $d>2p$, where $\mathcal{W}_p$ denotes the $p$-Wasserstein metric. An effective tool to combat this curse of dimensionality is the smooth Wasserstein distance $\mathcal{W}{(\sigma)}_p$, which measures the distance between two probability measures after having convolved them with isotropic Gaussian noise $\mathcal{N}(0,\sigma2\text{I})$. In this paper we apply this smoothing technique to the adapted Wasserstein distance. We show that the smooth adapted Wasserstein distance $\mathcal{A}\mathcal{W}_p{(\sigma)}$ achieves the fast rate of convergence $\mathbb{E}[\mathcal{A}\mathcal{W}_p{(\sigma)}(\hat{\mu}_n, \mu)]\lesssim n{-1/2}$, if $\mu$ is subgaussian. This result follows from the surprising fact, that any subgaussian measure $\mu$ convolved with a Gaussian distribution has locally Lipschitz kernels.

Summary

We haven't generated a summary for this paper yet.