Papers
Topics
Authors
Recent
Search
2000 character limit reached

A Statistical Analysis of Wasserstein Autoencoders for Intrinsically Low-dimensional Data

Published 24 Feb 2024 in cs.LG, math.ST, stat.ML, and stat.TH | (2402.15710v1)

Abstract: Variational Autoencoders (VAEs) have gained significant popularity among researchers as a powerful tool for understanding unknown distributions based on limited samples. This popularity stems partly from their impressive performance and partly from their ability to provide meaningful feature representations in the latent space. Wasserstein Autoencoders (WAEs), a variant of VAEs, aim to not only improve model efficiency but also interpretability. However, there has been limited focus on analyzing their statistical guarantees. The matter is further complicated by the fact that the data distributions to which WAEs are applied - such as natural images - are often presumed to possess an underlying low-dimensional structure within a high-dimensional feature space, which current theory does not adequately account for, rendering known bounds inefficient. To bridge the gap between the theory and practice of WAEs, in this paper, we show that WAEs can learn the data distributions when the network architectures are properly chosen. We show that the convergence rates of the expected excess risk in the number of samples for WAEs are independent of the high feature dimension, instead relying only on the intrinsic dimension of the data distribution.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (48)
  1. Neural network learning: Theoretical foundations. Cambridge University Press.
  2. Wasserstein generative adversarial networks. In International Conference on Machine Learning, pages 214–223. PMLR.
  3. Nearly-tight VC-dimension and pseudodimension bounds for piecewise linear neural networks. The Journal of Machine Learning Research, 20(1):2285–2301.
  4. Rademacher and Gaussian Complexities: Risk Bounds and Structural Results. Journal of Machine Learning Research, 3(Nov):463–482.
  5. Concentration inequalities: A nonasymptotic theory of independence. Oxford University Press.
  6. Statistical regeneration guarantees of the wasserstein autoencoder with latent space consistency. Advances in Neural Information Processing Systems, 34:17098–17110.
  7. On the statistical properties of generative adversarial models for low intrinsic data dimension. arXiv preprint arXiv:2401.15801.
  8. Distribution approximation and statistical estimation guarantees of generative adversarial networks. arXiv preprint arXiv:2002.03938.
  9. On deep generative models for approximation and estimation of distributions on manifolds. Advances in Neural Information Processing Systems, 35:10615–10628.
  10. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pages 248–255.
  11. Adversarial feature learning. In International Conference on Learning Representations.
  12. Automatic chemical design using a data-driven continuous representation of molecules. ACS central science, 4(2):268–276.
  13. Generative Adversarial Nets. In Advances in Neural Information Processing Systems, volume 27. Curran Associates, Inc.
  14. Draw: A recurrent neural network for image generation. In International Conference on Machine Learning, pages 1462–1471. PMLR.
  15. A kernel two-sample test. Journal of Machine Learning Research, 13(25):723–773.
  16. Improved Training of Wasserstein GANs. Advances in Neural Information Processing Systems, 30.
  17. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in Neural Information Processing Systems, 30.
  18. beta-vae: Learning basic visual concepts with a constrained variational framework. In International Conference on Learning Representations.
  19. An error analysis of generative adversarial networks for learning distributions. Journal of Machine Learning Research, 23(116):1–43.
  20. A primal-dual link between gans and autoencoders. In Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc.
  21. Karr, A. F. (1993). Probability. Springer Texts in Statistics. Springer New York, NY, 1 edition.
  22. Uniform convergence rate of the kernel density estimator adaptive to intrinsic volume dimension. In International Conference on Machine Learning, pages 3398–3407. PMLR.
  23. Adam: A method for stochastic optimization. International Conference on Learning Representations.
  24. Auto-encoding variational bayes. In International Conference on Learning Representations.
  25. Variational autoencoders in the presence of low-dimensional data: landscape and implicit bias. In International Conference on Learning Representations.
  26. ϵitalic-ϵ\epsilonitalic_ϵ-entropy and ϵitalic-ϵ\epsilonitalic_ϵ-capacity of sets in function spaces. Translations of the American Mathematical Society, 17:277–364.
  27. Liang, T. (2021). How well generative adversarial networks learn distributions. The Journal of Machine Learning Research, 22(1):10366–10406.
  28. Deep nonparametric estimation of intrinsic data structures by chart autoencoders: Generalization error and robustness. arXiv preprint arXiv:2303.09863.
  29. Non-asymptotic error bounds for bidirectional gans. Advances in Neural Information Processing Systems, 34:12328–12339.
  30. Adversarial autoencoders. In International Conference on Learning Representations.
  31. Adversarial variational bayes: Unifying variational autoencoders and generative adversarial networks. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 2391–2400. JMLR. org.
  32. Adaptive approximation and generalization of deep neural network with intrinsic dimensionality. Journal of Machine Learning Research, 21(174):1–38.
  33. The intrinsic dimension of images and its impact on learning. In International Conference on Learning Representations.
  34. Zero-shot text-to-image generation. In International Conference on Machine Learning, pages 8821–8831. PMLR.
  35. Variational autoencoders pursue pca directions (by accident). In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12406–12415.
  36. Hanson-Wright inequality and sub-gaussian concentration. Electronic Communications in Probability, 18(none):1 – 9.
  37. Statistical guarantees for generative models without domination. In Algorithmic Learning Theory, pages 1051–1071. PMLR.
  38. Learning structured output representation using deep conditional generative models. Advances in neural information processing systems, 28.
  39. Efficiently trainable text-to-speech system based on deep convolutional networks with guided attention. In 2018 IEEE international conference on acoustics, speech and signal processing (ICASSP), pages 4784–4788. IEEE.
  40. On empirical bayes variational autoencoder: An excess risk bound. In Conference on Learning Theory, pages 4068–4125. PMLR.
  41. Wasserstein auto-encoders. International Conference on Learning Representations.
  42. Neural discrete representation learning. Advances in Neural Information Processing Systems, 30.
  43. Vershynin, R. (2018). High-dimensional probability: An introduction with applications in data science, volume 47. Cambridge university press.
  44. Wainwright, M. J. (2019). High-dimensional statistics: A Non-asymptotic Viewpoint, volume 48. Cambridge University Press.
  45. Sharp asymptotic and finite-sample rates of convergence of empirical measures in Wasserstein distance. Bernoulli, 25(4A):2620 – 2648.
  46. Improved variational autoencoders for text modeling using dilated convolutions. In International Conference on Machine Learning, pages 3881–3890. PMLR.
  47. Yarotsky, D. (2017). Error bounds for approximations with deep relu networks. Neural Networks, 94:103–114.
  48. Infovae: Information maximizing variational autoencoders. arXiv preprint arXiv:1706.02262.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 5 likes about this paper.