Neural Diffusion Models (2310.08337v3)
Abstract: Diffusion models have shown remarkable performance on many generative tasks. Despite recent success, most diffusion models are restricted in that they only allow linear transformation of the data distribution. In contrast, broader family of transformations can potentially help train generative distributions more efficiently, simplifying the reverse process and closing the gap between the true negative log-likelihood and the variational approximation. In this paper, we present Neural Diffusion Models (NDMs), a generalization of conventional diffusion models that enables defining and learning time-dependent non-linear transformations of data. We show how to optimise NDMs using a variational bound in a simulation-free setting. Moreover, we derive a time-continuous formulation of NDMs, which allows fast and reliable inference using off-the-shelf numerical ODE and SDE solvers. Finally, we demonstrate the utility of NDMs with learnable transformations through experiments on standard image generation benchmarks, including CIFAR-10, downsampled versions of ImageNet and CelebA-HQ. NDMs outperform conventional diffusion models in terms of likelihood and produce high-quality samples.
- Building normalizing flows with stochastic interpolants. arXiv preprint arXiv:2209.15571, 2022.
- Neural ordinary differential equations. Advances in neural information processing systems, 31, 2018.
- Likelihood training of schr\\\backslash\” odinger bridge using forward-backward sdes theory. arXiv preprint arXiv:2110.11291, 2021.
- Generative adversarial networks: An overview. IEEE Signal Processing Magazine, 35(1):53–65, 2018.
- Good semi-supervised learning that requires a bad gan. Advances in neural information processing systems, 30, 2017.
- Soft diffusion: Score matching for general corruptions. arXiv preprint arXiv:2209.05442, 2022.
- Diffusion schrödinger bridge with applications to score-based generative modeling. Advances in Neural Information Processing Systems, 34:17695–17709, 2021.
- Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp. 248–255. Ieee, 2009.
- Deng, L. The mnist database of handwritten digit images for machine learning research [best of the web]. IEEE signal processing magazine, 29(6):141–142, 2012.
- Diffusion models beat GANs on image synthesis. arXiv preprint arXiv:2105.05233, 2021.
- A family of embedded runge-kutta formulae. Journal of computational and applied mathematics, 6(1):19–26, 1980.
- Generative adversarial nets. Advances in neural information processing systems, 27, 2014.
- Ffjord: Free-form continuous dynamics for scalable reversible generative models. arXiv preprint arXiv:1810.01367, 2018.
- f-dm: A multi-stage diffusion model via progressive signal transformation. arXiv preprint arXiv:2210.04955, 2022.
- Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017.
- Denoising diffusion probabilistic models. arXiv preprint arXiv:2006.11239, 2020.
- Blurring diffusion models. arXiv preprint arXiv:2209.05557, 2022.
- Progressive growing of gans for improved quality, stability, and variation. arXiv preprint arXiv:1710.10196, 2017.
- Maximum likelihood training of implicit nonlinear diffusion models. arXiv preprint arXiv:2205.13699, 2022.
- Auto-encoding variational Bayes. arXiv preprint arXiv:1312.6114, 2013.
- Variational diffusion models. arXiv preprint arXiv:2107.00630, 2, 2021.
- Learning multiple layers of features from tiny images. 2009.
- Minimizing trajectory curvature of ode-based generative models. arXiv preprint arXiv:2301.12003, 2023.
- Flow matching for generative modeling. arXiv preprint arXiv:2210.02747, 2022.
- Pseudo numerical methods for diffusion models on manifolds. arXiv preprint arXiv:2202.09778, 2022a.
- Liu, Q. Rectified flow: A marginal preserving approach to optimal transport. arXiv preprint arXiv:2209.14577, 2022.
- Flow straight and fast: Learning to generate and transfer data with rectified flow. arXiv preprint arXiv:2209.03003, 2022b.
- MacKay, D. J. Information theory, inference and learning algorithms. Cambridge university press, 2003.
- Action matching: A variational method for learning stochastic dynamics from samples. arXiv preprint arXiv:2210.06662, 2022.
- Improved denoising diffusion probabilistic models. arXiv preprint arXiv:2102.09672, 2021.
- Diffenc: Variational diffusion with a learned encoder. arXiv preprint arXiv:2310.19789, 2023.
- Stochastic differential equations. Springer, 2003.
- Normalizing flows for probabilistic modeling and inference. The Journal of Machine Learning Research, 22(1):2617–2680, 2021.
- Peluchetti, S. Non-denoising forward-time diffusions.
- Grad-tts: A diffusion probabilistic model for text-to-speech. In International Conference on Machine Learning, pp. 8599–8608. PMLR, 2021.
- Stochastic backpropagation and approximate inference in deep generative models. In International conference on machine learning, pp. 1278–1286. PMLR, 2014.
- Generative modelling with inverse heat dissipation. arXiv preprint arXiv:2206.13397, 2022.
- High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10684–10695, 2022.
- Image super-resolution via iterative refinement. arXiv preprint arXiv:2104.07636, 2021.
- Progressive distillation for fast sampling of diffusion models. arXiv preprint arXiv:2202.00512, 2022.
- Where to diffuse, how to diffuse, and how to get back: Automated learning for multivariate diffusions. arXiv preprint arXiv:2302.07261, 2023.
- Differential equations, dynamical systems, and linear algebra, volume 60. Elsevier, 1974.
- Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning, pp. 2256–2265. PMLR, 2015.
- Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502, 2020a.
- Pixeldefend: Leveraging generative models to understand and defend against adversarial examples. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=rJUYGxbCW.
- Score-based generative modeling through stochastic differential equations. arXiv preprint arXiv:2011.13456, 2020b.
- Maximum likelihood training of score-based diffusion models. Advances in Neural Information Processing Systems, 34:1415–1428, 2021.
- It\\\backslash\^{{\{{o}}\}}-taylor sampling scheme for denoising diffusion probabilistic models using ideal derivatives. arXiv preprint arXiv:2112.13339, 2021.
- Tomczak, J. M. Deep generative modeling. Springer, 2022.
- Diffusion probabilistic modeling of protein backbones in 3D for the motif-scaffolding problem. In The Eleventh International Conference on Learning Representations, 2023.
- Score-based generative modeling in latent space. Advances in Neural Information Processing Systems, 34:11287–11302, 2021.
- Deep generative learning via schrödinger bridge. In International Conference on Machine Learning, pp. 10794–10804. PMLR, 2021.
- Broadly applicable and accurate protein design by integrating structure prediction networks and diffusion generative models. bioRxiv, pp. 2022–12, 2022.
- Practical and asymptotically exact conditional sampling in diffusion models. arXiv preprint arXiv:2306.17775, 2023.
- Tackling the generative learning trilemma with denoising diffusion GANs. arXiv preprint arXiv:2112.07804, 2021.
- Diffusion models: A comprehensive survey of methods and applications. arXiv preprint arXiv:2209.00796, 2022.