Empirical Wasserstein Distance
- Empirical Wasserstein distance is a metric that measures discrepancies between probability distributions using optimal transport techniques on both discrete and continuous spaces.
- Recent research establishes explicit distributional theorems, finite-sample rates, and consistent bootstrap methodologies to support rigorous statistical inference.
- Its practical applications span fingerprint verification, metagenomic analysis, and distributional robustness, making it a valuable tool in high-dimensional data analysis.
The empirical Wasserstein distance quantifies the discrepancy between probability distributions using observed data, forming a pivotal object for statistical inference, quantization, probabilistic modeling, machine learning, and distributional robustness. On a finite or discrete sample space, as well as for general continuous distributions, its definition, limiting behavior, and inferential framework rely on optimal transport theory and the structure of empirical processes. Recent research has established explicit distributional theorems, finite-sample rates, confidence interval constructions, and connections between empirical Wasserstein distances and linear programming, Hadamard differentiability, and alternative resampling methods.
1. Definition and Linear Programming Structure
On a finite metric space with metric , any probability measure is a mass vector , , . The empirical Wasserstein distance of order between two vectors is
where is the set of joint distributions with prescribed marginals: , , .
This optimization is formulated as a linear program (LP), whose dual exposes the intrinsic geometry of Wasserstein balls:
- Primal (transport plan): minimizes the total transport cost under marginal constraints.
- Dual: maximizes the affine objective over dual variables , subject to .
Solution sets contain all optimal dual pairs and are critical for sensitivity analysis and limit distribution formulations (Sommerfeld et al., 2016).
2. Asymptotic Distribution and Limit Theorems
Suppose samples i.i.d. from , i.i.d. from , giving empirical laws . Define multinomial covariance matrices and centered Gaussian random vectors: The main theorem (Sommerfeld et al., 2016) characterizes weak convergence for all :
- One-sample null ( known):
- One-sample alternative ( known):
- Two-sample null ():
- Two-sample alternative: , .
This framework generalizes classical Donsker theory and reveals the optimal linear programming structure behind Wasserstein empirical processes.
3. Statistical Inference and Bootstrap Methodology
The non-linear, supreme-type Hadamard differentiability of (directional, but not full Fréchet) demands specialized approaches to produce valid inferential results:
- Classical ("n-out-of-n") bootstrap fails: It does not capture the limiting law since the local empirical process is not sufficiently approximated (Sommerfeld et al., 2016).
- Consistent alternatives:
- m-out-of-n bootstrap: Selects subsamples (with ), consistent but involves additional tuning.
- Directional bootstrap (Fang–Santos): Uses the empirical process vector, plugged into the explicit Hadamard derivative, yielding consistent inference.
This theoretical apparatus permits the construction of confidence intervals and hypothesis tests for empirical Wasserstein distances in various regimes.
4. Finite-Sample Confidence Regions and Quantile Calculations
Explicit formulas for finite sample inference are developed as follows (Sommerfeld et al., 2016, Cohen et al., 2019):
- For null hypotheses (), compute quantiles of the limit law via simulation:
- Estimate population vectors and covariances from .
- Generate Gaussian samples for , solve the dual LP for each replicate, and power-transform as needed.
- Empirical quantiles determine confidence intervals.
For one-dimensional cases, tight bounds on the normalized quantiles have been established (functional CLT; Brownian bridge), and explicit confidence regions are constructed via extremal distributions—mixtures of two-point and uniform distributions on the unit interval (Cohen et al., 2019). Extensions to higher dimensions are conjectured, supported by numerical evidence.
5. Examples and Applications
Empirical Wasserstein distance underpins applied analyses across multiple domains:
- Fingerprint Verification: Empirical minutiae histograms compared via Wasserstein provide significant differentiations between real and synthetic datasets. Optimal transport plans visualize mass shifts among feature bins.
- Metagenomic Communities: Wasserstein-based pairwise sample distances, along with confidence intervals, quantitatively distinguish intra- and inter-personal differences in gut microbiome composition.
- Distributional Robustness and Goodness-of-Fit: The limit laws for empirical Wasserstein distances provide finite-sample guarantees for robust optimization, model validation, and two-sample or one-sample tests in generative and biological modeling.
- Quantization: The empirical Wasserstein distance matches optimal rates of finite quantization in Euclidean and infinite-dimensional Gaussian spaces, demonstrating rate-optimality for empirical processes (Boissard et al., 2011).
6. Technical Underpinnings: Hadamard Differentiability and LP Sensitivity
Central to the limiting theory is the directional Hadamard differentiability of the Wasserstein cost functional on finite spaces. Sensitivity analysis of parametric LP yields the precise form of the derivative: This sensitivity analysis enables direct application of the functional delta method for limiting distributions. For continuous underlying distributions (e.g., Gaussian), explicit closed-form derivatives and CLTs for the empirical Wasserstein distance have been derived, leveraging Spectral and Fréchet differentiability (Rippl et al., 2015).
7. Computational Implementation and Moderate Sample Validity
All inferential statements and confidence region calculations can be performed using standard LP solvers and multivariate normal samplers. Theoretical quantile formulas and simulation-based approximations support valid inference even for moderate sample sizes, with empirical evidence demonstrating the effectiveness and sharpness of derived finite-sample intervals (Sommerfeld et al., 2016).
In summary, the empirical Wasserstein distance is characterized, on finite metric spaces, by an explicit LP-optimization structure. Its distributional and inferential properties are governed by the geometry of the transport polyhedron and the Hadamard differentiability of the transport cost. The combination of theoretical insights, dual LP sensitivity, and computational tractability enables rigorous hypothesis testing, confidence interval construction, and broad application in complex, high-dimensional data analysis.