Deep MMD Gradient Flow without adversarial training (2405.06780v1)
Abstract: We propose a gradient flow procedure for generative modeling by transporting particles from an initial source distribution to a target distribution, where the gradient field on the particles is given by a noise-adaptive Wasserstein Gradient of the Maximum Mean Discrepancy (MMD). The noise-adaptive MMD is trained on data distributions corrupted by increasing levels of noise, obtained via a forward diffusion process, as commonly used in denoising diffusion probabilistic models. The result is a generalization of MMD Gradient Flow, which we call Diffusion-MMD-Gradient Flow or DMMD. The divergence training procedure is related to discriminator training in Generative Adversarial Networks (GAN), but does not require adversarial training. We obtain competitive empirical performance in unconditional image generation on CIFAR10, MNIST, CELEB-A (64 x64) and LSUN Church (64 x 64). Furthermore, we demonstrate the validity of the approach when MMD is replaced by a lower bound on the KL divergence.
- Fast inference in denoising diffusion models via mmd finetuning, 2023.
- Neural wasserstein gradient flows for maximum mean discrepancies with riesz kernels, 2023.
- Gradient Flows in Metric Spaces and in the Space of Probability Measures. Lectures in Mathematics ETH Zürich. Birkhäuser, 2. ed edition, 2008. ISBN 978-3-7643-8722-8 978-3-7643-8721-1. OCLC: 254181287.
- Refining deep generative models via discriminator gradient flow, 2021.
- On gradient regularizers for mmd gans. Advances in neural information processing systems, 31, 2018.
- Maximum mean discrepancy gradient flow, 2019.
- Generalized energy based models, 2021.
- Wasserstein gan, 2017.
- Generalization and equilibrium in generative adversarial nets (gans), 2017.
- Demystifying mmd gans, 2021.
- Bortoli, V. D. Convergence of denoising diffusion models under the manifold hypothesis, 2023.
- Large scale gan training for high fidelity natural image synthesis, 2019.
- The union of manifolds hypothesis and its implications for deep generative modelling. arXiv preprint arXiv:2207.02862, 2022.
- Your gan is secretly an energy-based model and you should use discriminator driven latent sampling, 2021.
- Sampling is as easy as learning the score: theory for diffusion models with minimal data assumptions. arXiv preprint arXiv:2209.11215, 2022.
- Variational wasserstein gradient flow, 2022.
- Testing the manifold hypothesis. Journal of the American Mathematical Society, 29(4):983–1049, 2016.
- A neural tangent kernel perspective of gans, 2022.
- Unifying gans and score-based diffusion as generative particle models, 2023.
- Learning generative models with sinkhorn divergences. In Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics, volume 84 of Proceedings of Machine Learning Research, pp. 1608–1617. PMLR, 2018.
- Kale flow: A relaxed kl gradient flow for probabilities with disjoint support, 2021.
- Generative adversarial networks, 2014.
- A kernel two-sample test. Journal of Machine Learning Research, 13(25):723–773, 2012. URL http://jmlr.org/papers/v13/gretton12a.html.
- Improved training of wasserstein gans, 2017.
- Posterior sampling based on gradient flows of the mmd with negative distance kernel, 2023.
- Deep generative wasserstein gradient flows, 2023. URL https://openreview.net/forum?id=zjSeBTEdXp1.
- Generative sliced mmd flows with riesz kernels, 2023.
- Gans trained by a two time-scale update rule converge to a local nash equilibrium, 2018.
- Denoising diffusion probabilistic models, 2020.
- Imagen video: High definition video generation with diffusion models. arXiv preprint arXiv:2210.02303, 2022.
- Hyvärinen, A. Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(24):695–709, 2005. URL http://jmlr.org/papers/v6/hyvarinen05a.html.
- Neural tangent kernel: Convergence and generalization in neural networks, 2020.
- The variational formulation of the fokker–planck equation. SIAM Journal on Mathematical Analysis, 29(1):1–17, 1998. doi: 10.1137/S0036141096303359. URL https://doi.org/10.1137/S0036141096303359.
- Training generative adversarial networks with limited data, 2020a.
- Analyzing and improving the image quality of stylegan, 2020b.
- Adam: A method for stochastic optimization, 2017.
- On convergence and stability of gans, 2017.
- Learning multiple layers of features from tiny images. 2009.
- Voicebox: Text-guided multilingual universal speech generation at scale. arXiv preprint arXiv:2306.15687, 2023.
- Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998. doi: 10.1109/5.726791.
- Mmd gan: Towards deeper understanding of moment matching network, 2017.
- On error propagation of diffusion models, 2024.
- Generative adversarial networks for image and video synthesis: Algorithms and applications, 2020.
- Deep learning face attributes in the wild, 2015.
- Which training methods for gans do actually converge? In International conference on machine learning, pp. 3481–3490. PMLR, 2018.
- Muller, A. Integral probability metrics and their generating classes of functions. volume 29, pp. 429–443. Advances in Applied Probability, 1997.
- f-gan: Training generative neural samplers using variational divergence minimization, 2016.
- Pidstrigach, J. Score-based generative models detect manifolds. Advances in Neural Information Processing Systems, 35:35852–35865, 2022.
- Dreamfusion: Text-to-3d using 2d diffusion. arXiv preprint arXiv:2209.14988, 2022.
- High-resolution image synthesis with latent diffusion models, 2022.
- U-net: Convolutional networks for biomedical image segmentation, 2015.
- Photorealistic text-to-image diffusion models with deep language understanding. Advances in Neural Information Processing Systems, 35:36479–36494, 2022.
- Improved techniques for training gans, 2016.
- Santambrogio, F. Optimal transport for applied mathematicians. Birkäuser, NY, 55(58-63):94, 2015.
- Adversarial diffusion distillation. arXiv preprint arXiv:2311.17042, 2023.
- Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond. The MIT Press, 06 2018. ISBN 9780262256933. doi: 10.7551/mitpress/4175.001.0001. URL https://doi.org/10.7551/mitpress/4175.001.0001.
- Deep unsupervised learning using nonequilibrium thermodynamics, 2015.
- Score-based generative modeling through stochastic differential equations. arXiv preprint arXiv:2011.13456, 2020.
- Consistency models. arXiv preprint arXiv:2303.01469, 2023.
- Going deeper with convolutions, 2014.
- A global geometric framework for nonlinear dimensionality reduction. science, 290(5500):2319–2323, 2000.
- Villani, C. Optimal Transport: Old and New. Grundlehren der mathematischen Wissenschaften. Springer Berlin Heidelberg, 2008. ISBN 9783540710509. URL https://books.google.co.uk/books?id=hV8o5R7_5tkC.
- Vincent, P. A connection between score matching and denoising autoencoders. Neural Computation, 23(7):1661–1674, 2011.
- Broadly applicable and accurate protein design by integrating structure prediction networks and diffusion generative models. BioRxiv, pp. 2022–12, 2022.
- Tackling the generative learning trilemma with denoising diffusion gans, 2022.
- Ufogen: You forward once large scale text-to-image generation via diffusion gans. arXiv preprint arXiv:2311.09257, 2023.
- Eliminating lipschitz singularities in diffusion models, 2023.
- Lsun: Construction of a large-scale image dataset using deep learning with humans in the loop, 2016.
- On the discrimination-generalization tradeoff in gans. In 6th International Conference on Learning Representations, 2018.