Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
184 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Note on the Convergence of Denoising Diffusion Probabilistic Models (2312.05989v2)

Published 10 Dec 2023 in cs.LG

Abstract: Diffusion models are one of the most important families of deep generative models. In this note, we derive a quantitative upper bound on the Wasserstein distance between the data-generating distribution and the distribution learned by a diffusion model. Unlike previous works in this field, our result does not make assumptions on the learned score function. Moreover, our bound holds for arbitrary data-generating distributions on bounded instance spaces, even those without a density w.r.t. the Lebesgue measure, and the upper bound does not suffer from exponential dependencies. Our main result builds upon the recent work of Mbacke et al. (2023) and our proofs are elementary.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (33)
  1. Generative modeling with denoising auto-encoders and langevin sampling. arXiv preprint arXiv:2002.00107, 2020.
  2. Wavegrad: Estimating gradients for waveform generation. In International Conference on Learning Representations, 2021.
  3. Sampling is as easy as learning the score: theory for diffusion models with minimal data assumptions. In The Eleventh International Conference on Learning Representations, 2023.
  4. Valentin De Bortoli. Convergence of denoising diffusion models under the manifold hypothesis. Transactions on Machine Learning Research, 2022. ISSN 2835-8856.
  5. Diffusion schrödinger bridge with applications to score-based generative modeling. Advances in Neural Information Processing Systems, 34:17695–17709, 2021.
  6. Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 34:8780–8794, 2021.
  7. Testing the manifold hypothesis. Journal of the American Mathematical Society, 29(4):983–1049, 2016.
  8. Generative adversarial nets. In Advances in Neural Information Processing Systems, volume 27, 2014.
  9. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840–6851, 2020.
  10. Cascaded diffusion models for high fidelity image generation. The Journal of Machine Learning Research, 23(1):2249–2281, 2022.
  11. Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(4), 2005.
  12. Diederik P. Kingma and M. Welling. Auto-encoding variational Bayes. CoRR, abs/1312.6114, 2014.
  13. Score-based generative modeling secretly minimizes the wasserstein distance. Advances in Neural Information Processing Systems, 35:20205–20217, 2022.
  14. Convergence for score-based generative modeling with polynomial complexity. Advances in Neural Information Processing Systems, 35:22870–22882, 2022.
  15. Convergence of score-based generative modeling for general data distributions. In International Conference on Algorithmic Learning Theory, pp.  946–985. PMLR, 2023.
  16. Towards faster non-asymptotic convergence for diffusion-based generative models. arXiv preprint arXiv:2306.09251, 2023.
  17. Srdiff: Single image super-resolution with diffusion probabilistic models. Neurocomputing, 479:47–59, 2022.
  18. Calvin Luo. Understanding diffusion models: A unified perspective. arXiv preprint arXiv:2208.11970, 2022.
  19. Statistical guarantees for variational autoencoders using PAC-bayesian theory. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  20. Sample complexity of testing the manifold hypothesis. Advances in neural information processing systems, 23, 2010.
  21. Jakiw Pidstrigach. Score-based generative models detect manifolds. Advances in Neural Information Processing Systems, 35:35852–35865, 2022.
  22. The intrinsic dimension of images and its impact on learning. In International Conference on Learning Representations, 2021.
  23. Grad-tts: A diffusion probabilistic model for text-to-speech. In International Conference on Machine Learning, pp.  8599–8608. PMLR, 2021.
  24. Stochastic backpropagation and approximate inference in deep generative models. In International conference on machine learning, pp.  1278–1286. PMLR, 2014.
  25. Unit-ddpm: Unpaired image translation with denoising diffusion probabilistic models. arXiv preprint arXiv:2104.05358, 2021.
  26. Deep unsupervised learning using nonequilibrium thermodynamics. In International conference on machine learning, pp.  2256–2265. PMLR, 2015.
  27. Generative modeling by estimating gradients of the data distribution. Advances in neural information processing systems, 32, 2019.
  28. Maximum likelihood training of score-based diffusion models. Advances in Neural Information Processing Systems, 34:1415–1428, 2021a.
  29. Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021b.
  30. Effective data augmentation with diffusion models. arXiv preprint arXiv:2302.07944, 2023.
  31. Cédric Villani. Optimal transport: old and new, volume 338. Springer, 2009.
  32. Pascal Vincent. A connection between score matching and denoising autoencoders. Neural computation, 23(7):1661–1674, 2011.
  33. 3d shape generation and completion through point-voxel diffusion. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.  5826–5835, 2021.
Citations (3)

Summary

We haven't generated a summary for this paper yet.