Papers
Topics
Authors
Recent
Search
2000 character limit reached

Statistical, Robustness, and Computational Guarantees for Sliced Wasserstein Distances

Published 17 Oct 2022 in stat.ML and cs.LG | (2210.09160v1)

Abstract: Sliced Wasserstein distances preserve properties of classic Wasserstein distances while being more scalable for computation and estimation in high dimensions. The goal of this work is to quantify this scalability from three key aspects: (i) empirical convergence rates; (ii) robustness to data contamination; and (iii) efficient computational methods. For empirical convergence, we derive fast rates with explicit dependence of constants on dimension, subject to log-concavity of the population distributions. For robustness, we characterize minimax optimal, dimension-free robust estimation risks, and show an equivalence between robust sliced 1-Wasserstein estimation and robust mean estimation. This enables lifting statistical and algorithmic guarantees available for the latter to the sliced 1-Wasserstein setting. Moving on to computational aspects, we analyze the Monte Carlo estimator for the average-sliced distance, demonstrating that larger dimension can result in faster convergence of the numerical integration error. For the max-sliced distance, we focus on a subgradient-based local optimization algorithm that is frequently used in practice, albeit without formal guarantees, and establish an $O(ε{-4})$ computational complexity bound for it. Our theory is validated by numerical experiments, which altogether provide a comprehensive quantitative account of the scalability question.

Citations (30)

Summary

  • The paper establishes convergence rates, robustness, and computational guarantees for Sliced Wasserstein distances in high-dimensional spaces.
  • It shows that empirical convergence adapts to sample size and effective dimensionality while maintaining dimension-free robust estimation risks.
  • The authors propose efficient computational methods, including Monte Carlo integration and subgradient optimization, to overcome non-convexity challenges.

Statistical, Robustness, and Computational Guarantees for Sliced Wasserstein Distances

Introduction

"Sliced Wasserstein distances" offer a potential solution to the computational challenges faced by the classic Wasserstein distances, especially in high-dimensional spaces. This paper aims to quantify the benefits of using Sliced Wasserstein distances from three perspectives: statistical convergence, robustness to data contamination, and computational efficiency. The sliced version of Wasserstein distances considered here are defined based on the average and maximum Wasserstein distances over one-dimensional projections of high-dimensional distributions. The paper derives explicit empirical convergence rates with dependencies on dimensionality, provides robust estimation risks that are dimension-independent, and furnishes computational methods with formal guarantees for efficient calculation.

Empirical Convergence

The paper provides convergence rates for the empirical Sliced Wasserstein distances that reveal their scalability in high-dimensional settings. These distances exploit projections to lower dimensions, which allows them to mitigate the usual dimensionality curse encountered with classical Wasserstein distances.

For distributions that are log-concave, it is shown that both average- and max-sliced Wasserstein distances exhibit convergence rates dependent on the sample size and the effective dimensionality (rank) of the data's covariance matrix. Specifically, the distance decays as O(n1/max{2,p})\mathcal{O}(n^{-1/\max\{2,p\}}) given nn samples, adapting well to the structure of the data. Figure 1

Figure 1

Figure 1

Figure 1

Figure 1: Depiction of absolute deviation of empirical Sliced Wasserstein from true distance under Model (1).

Robustness to Data Contamination

One key advantage of Sliced Wasserstein distances is their robustness to data contamination. The paper demonstrates that these distances admit minimax optimal risk bounds for robust estimation that are dimension-free. Specifically, the worst-case estimation risk over a class of perturbed data decreases without dependence on dimensional factors that commonly impair other metrics like the classic Wasserstein distance.

The results leverage a notion of equivalence in robustness between one-dimensional slicing and robust mean estimation, which facilitates lifting bounds from known mean estimation results to the Wasserstein metrics considered. Figure 2

Figure 2

Figure 2: Error of Sliced Wasserstein estimation under Model (3) where data contamination exists.

Computational Guarantees

For computational efficiency, the authors analyze methods to rapidly compute the Sliced Wasserstein distances. The Monte Carlo integration method is proven effective for computing the average-sliced Wasserstein distance with an error that can diminish as dimensionality increases. This result defies the typical curse of dimensionality by showing that the accuracy of this straightforward estimator can actually benefit from increasing dimension in certain cases.

For maximum-slicing, a subgradient method for local optimization is shown to have effective computational complexity, including concrete bounds on the number of iterations required to reach a stationary point. Global optimization strategies based on the LIPO algorithm were also benchmarked, highlighting how even those methods can achieve guaranteed performance in practical scenarios despite non-convexity challenges. Figure 3

Figure 3

Figure 3: Illustration of robust estimation errors compared between robust estimation using MSW1 and classic W1, highlighting the dimension-free property of sliced distances.

Conclusion

Sliced Wasserstein distances provide substantial scalability advantages over the classic Wasserstein distance in the context of high-dimensional applications. The adaptation to the structure of data, robustness to outliers, and computational efficiency achieved by slicing highlight their utility in modern machine learning tasks. Future research directions could include extending the analysis to different forms of sliced distances or exploring alternative computational strategies to further enhance performance and applicability.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.