Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Neural Implicit Manifold Learning for Topology-Aware Density Estimation (2206.11267v2)

Published 22 Jun 2022 in stat.ML and cs.LG

Abstract: Natural data observed in $\mathbb{R}n$ is often constrained to an $m$-dimensional manifold $\mathcal{M}$, where $m < n$. This work focuses on the task of building theoretically principled generative models for such data. Current generative models learn $\mathcal{M}$ by mapping an $m$-dimensional latent variable through a neural network $f_\theta: \mathbb{R}m \to \mathbb{R}n$. These procedures, which we call pushforward models, incur a straightforward limitation: manifolds cannot in general be represented with a single parameterization, meaning that attempts to do so will incur either computational instability or the inability to learn probability densities within the manifold. To remedy this problem, we propose to model $\mathcal{M}$ as a neural implicit manifold: the set of zeros of a neural network. We then learn the probability density within $\mathcal{M}$ with a constrained energy-based model, which employs a constrained variant of Langevin dynamics to train and sample from the learned manifold. In experiments on synthetic and natural data, we show that our model can learn manifold-supported distributions with complex topologies more accurately than pushforward models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (109)
  1. Sine-skewed toroidal distributions and their application in protein bioinformatics. Biostatistics, 23(3):685–704, 2020.
  2. Generalized energy based models. In International Conference on Learning Representations, 2021.
  3. Wasserstein generative adversarial networks. In International Conference on Machine Learning, pp. 214–223, 2017.
  4. Understanding and mitigating exploding inverses in invertible neural networks. In International Conference on Artificial Intelligence and Statistics, pp.  1792–1800, 2021.
  5. Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8):1798–1828, 2013.
  6. Pyro: Deep universal probabilistic programming. Journal of Machine Learning Research, 20(28):1–6, 2019.
  7. A generative, probabilistic model of local protein structure. Proceedings of the National Academy of Sciences, 105(26):8932–8937, 2008.
  8. G Robert Brakenridge. Global active archive of large flood events. Dartmouth Flood Observatory, University of Colorado, 2010.
  9. Flows for simultaneous manifold learning and density estimation. In Advances in Neural Information Processing Systems, 2020.
  10. Verifying the union of manifolds hypothesis for image data. In International Conference on Learning Representations, 2023.
  11. A family of MCMC methods on implicitly defined manifolds. In International Conference on Artificial Intelligence and Statistics, pp.  161–172, 2012.
  12. A limited memory algorithm for bound constrained optimization. SIAM Journal on Scientific Computing, 16(5):1190–1208, 1995.
  13. Geodesic Monte Carlo on embedded manifolds. Scandinavian Journal of Statistics, 40(4):825–845, 2013.
  14. Theophilos Cacoullos. Estimation of a multivariate density. Annals of the Institute of Statistical Mathematics, 18(1):179–189, 1966.
  15. On the local behavior of spaces of natural images. International Journal of Computer Vision, 76(1):1–12, 2008.
  16. Entropic issues in likelihood-based OOD detection. In I (Still) Can’t Believe It’s Not Better! Workshop at NeurIPS 2021, pp.  21–26, 2021.
  17. Rectangular flows for manifold learning. In Advances in Neural Information Processing Systems, 2021.
  18. Your GAN is secretly an energy-based model and you should use discriminator driven latent sampling. In Advances in Neural Information Processing Systems, 2020.
  19. Neural ordinary differential equations. In Advances in Neural Information Processing Systems, 2018.
  20. Learning implicit fields for generative shape modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  5939–5948, 2019.
  21. Relaxing bijectivity constraints with continuously indexed normalising flows. In International Conference on Machine Learning, pp. 2133–2143, 2020.
  22. Caloman: Fast generation of calorimeter showers with density estimation on learned manifolds. arXiv preprint arXiv:2211.15380, 2022.
  23. Balázs Csanád Csáji. Approximation with artificial neural networks. Faculty of Sciences, Eötvös Loránd University, Hungary, 24(48):7, 2001.
  24. Diagnosing and enhancing VAE models. In International Conference on Learning Representations, 2019.
  25. Riemannian score-based generative modeling. In Advances in Neural Information Processing Systems, 2022.
  26. Sampling from a manifold. In Advances in Modern Statistical Theory and Applications: A Festschrift in Honor of Morris L. Eaton, pp.  102–125. Institute of Mathematical Statistics, 2013.
  27. Density estimation using Real NVP. In International Conference on Learning Representations, 2017.
  28. Implicit generation and modeling with energy based models. In Advances in Neural Information Processing Systems, 2019.
  29. Compositional visual generation with energy based models. In Advances in Neural Information Processing Systems, 2020.
  30. POT: Python optimal transport. Journal of Machine Learning Research, 22(78):1–8, 2021.
  31. GPyTorch: Blackbox matrix-matrix Gaussian process inference with GPU acceleration. In Advances in neural information processing systems, 2018.
  32. Normalizing flows on Riemannian manifolds. arXiv:1611.02304, 2016.
  33. From variational to deterministic autoencoders. In International Conference on Learning Representations, 2020.
  34. Josiah Willard Gibbs. Elementary Principles in Statistical Mechanics: Developed with Special Reference to the Rational Foundations of Thermodynamics. C. Scribner’s Sons, 1902.
  35. Generative adversarial nets. In Advances in Neural Information Processing Systems, 2014.
  36. Your classifier is secretly an energy based model and you should treat it like one. In International Conference on Learning Representations, 2020a.
  37. Learning the stein discrepancy for training and evaluating energy-based models without sampling. In International Conference on Machine Learning, pp. 3732–3747, 2020b.
  38. Implicit geometric regularization for learning shapes. In International Conference on Machine Learning, pp. 3789–3799, 2020.
  39. functorch: JAX-like composable function transforms for PyTorch. https://github.com/pytorch/functorch, 2021.
  40. Gaussian error linear units (GELUs). arXiv:1606.08415, 2016.
  41. GANs trained by a two time-scale update rule converge to a local Nash equilibrium. In Advances in Neural Information Processing Systems, 2017.
  42. beta-VAE: Learning basic visual concepts with a constrained variational framework. In International Conference on Learning Representations, 2017.
  43. Geoffrey E Hinton. Training products of experts by minimizing contrastive divergence. Neural Computation, 14(8):1771–1800, 2002.
  44. Denoising diffusion probabilistic models. In Advances in Neural Information Processing Systems, 2020.
  45. Multilayer feedforward networks are universal approximators. Neural Networks, 2(5):359–366, 1989.
  46. Aapo Hyvärinen. Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(24):695–709, 2005.
  47. On the need for topology-aware generative models for manifold-based defenses. In International Conference on Learning Representations, 2020.
  48. labml.ai annotated paper implementations, 2020.
  49. Density estimation on smooth manifolds with normalizing flows. arXiv:2106.03500, 2021.
  50. Adam: A method for stochastic optimization. In International Conference on Learning Representations, 2015.
  51. Auto-encoding variational Bayes. In International Conference on Learning Representations, 2014.
  52. Why normalizing flows fail to detect out-of-distribution data. In Advances in Neural Information Processing Systems, 2020.
  53. Trumpets: Injective flows for inference and inverse problems. In Uncertainty in Artificial Intelligence, pp.  1269–1278, 2021.
  54. Learning multiple layers of features from tiny images. Technical report, University of Toronto, 2009.
  55. Regularized autoencoders via relaxed injective probability flow. In International Conference on Artificial Intelligence and Statistics, pp.  4292–4301, 2020.
  56. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998.
  57. A tutorial on energy-based learning. Predicting Structured Data, 2006.
  58. John M Lee. Introduction to Smooth Manifolds. Springer New York, 2012.
  59. Diagnosing and fixing manifold overfitting in deep generative models. Transactions on Machine Learning Research, 2022.
  60. Adversarial autoencoders. arXiv:1511.05644, 2015.
  61. Protein bioinformatics and mixtures of bivariate von Mises distributions for angular data. Biometrics, 63(2):505–512, 2007.
  62. Riemannian continuous normalizing flows. In Advances in Neural Information Processing Systems, 2020.
  63. Disentangling disentanglement in variational autoencoders. In International Conference on Machine Learning, pp. 4402–4412, 2019.
  64. UMAP: Uniform manifold approximation and projection for dimension reduction. Journal of Open Source Software, 3(29):861, 2018.
  65. Occupancy networks: Learning 3D reconstruction in function space. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  4460–4470, 2019.
  66. Learning nonlinear constraints with contrastive backpropagation. In Proceedings of the IEEE International Joint Conference on Neural Networks, pp.  1302–1307, 2005.
  67. Learning in implicit generative models. arXiv:1610.03483, 2016.
  68. Do deep generative models know what they don’t know? In International Conference on Learning Representations, 2019.
  69. GIRAFFE: Representing scenes as compositional generative neural feature fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  11453–11464, 2021.
  70. Differentiable volumetric rendering: Learning implicit 3D representations without 3D supervision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  3504–3515, 2020.
  71. Learning non-convergent non-persistent short-run MCMC toward energy-based model. In Advances in Neural Information Processing Systems, 2019.
  72. On the anatomy of MCMC-based maximum likelihood learning of energy-based models. In Proceedings of the AAAI Conference on Artificial Intelligence, pp.  5272–5280, 2020.
  73. Numerical Optimization. Springer Science & Business Media, 2006.
  74. Submanifold density estimation. In Advances in Neural Information Processing Systems, 2009.
  75. DeepSDF: Learning continuous signed distance functions for shape representation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  165–174, 2019.
  76. PyTorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems, 2019.
  77. Xavier Pennec. Probabilities and statistics on Riemannian manifolds: Basic tools for geometric measurements. In Workshop on Nonlinear Signal and Image Processing, pp. 194–198, 1999.
  78. Composable effects for flexible and accelerated probabilistic programming in NumPyro. arXiv preprint arXiv:1912.11554, 2019.
  79. Jakiw Pidstrigach. Score-based generative models detect manifolds. In Advances in Neural Information Processing Systems, 2022.
  80. A survey of manifold learning for images. IPSJ Transactions on Computer Vision and Applications, 1:83–94, 2009.
  81. The intrinsic dimension of images and its impact on learning. In International Conference on Learning Representations, 2021.
  82. Bias-free scalable Gaussian processes via randomized truncations. In International Conference on Machine Learning, 2021.
  83. Searching for activation functions. arXiv:1710.05941, 2017.
  84. Variational inference with normalizing flows. In International Conference on Machine Learning, pp. 1530–1538, 2015.
  85. Stochastic backpropagation and approximate inference in deep generative models. In International Conference on Machine Learning, pp. 1278–1286, 2014.
  86. Normalizing flows on tori and spheres. In International Conference on Machine Learning, pp. 8083–8092, 2020.
  87. The manifold tangent classifier. In Advances in Neural Information Processing Systems, 2011.
  88. A stochastic approximation method. The Annals of Mathematical Statistics, 22(3):400–407, 1951.
  89. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  10684–10695, 2022.
  90. Tractable density estimation on learned manifolds with conformal embedding flows. In Advances in Neural Information Processing Systems, 2021.
  91. Walter Rudin. Real and Complex Analysis. McGraw-Hill, Inc., third edition, 1987.
  92. Can push-forward generative models fit multimodal distributions? In Advances in Neural Information Processing Systems, 2022.
  93. VQ-flows: Vector quantized local normalizing flows. In Uncertainty in Artificial Intelligence, pp.  1835–1845, 2022.
  94. Probabilistic model for two dependent circular variables. Biometrika, 89(3):719–723, 2002.
  95. Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning, pp. 2256–2265, 2015.
  96. Denoising diffusion implicit models. In International Conference on Learning Representations, 2021a.
  97. How to train your energy-based models. arXiv:2101.03288, 2021.
  98. Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021b.
  99. Exposing flaws of generative model evaluation metrics and their unfair treatment of diffusion models. In Advances in Neural Information Processing Systems, 2023.
  100. Wasserstein auto-encoders. In International Conference on Learning Representations, 2018.
  101. Bayesian learning via stochastic gradient Langevin dynamics. In International Conference on Machine Learning, pp. 681–688, 2011.
  102. Fashion-MNIST: A novel image dataset for benchmarking machine learning algorithms. arXiv:1708.07747, 2017.
  103. Generative latent flow. arXiv:1905.10485, 2019.
  104. VAEBM: A symbiosis between variational autoencoders and energy-based models. In International Conference on Learning Representations, 2021.
  105. A theory of generative ConvNet. In International Conference on Machine Learning, pp. 2635–2644, 2016.
  106. Tatsuya Yatagawa. torchmcubes: marching cubes for pytorch, 2021. URL https://github.com/tatsy/torchmcubes.
  107. Autoencoding under normalization constraints. In International Conference on Machine Learning, pp. 12087–12097, 2021.
  108. Monte Carlo on manifolds: Sampling densities and integrating functions. Communications on Pure and Applied Mathematics, 71(12):2609–2647, 2018.
  109. Learning manifold dimensions with conditional variational autoencoders. In Advances in Neural Information Processing Systems, 2022.
Citations (2)

Summary

We haven't generated a summary for this paper yet.