Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Observation-Guided Diffusion Probabilistic Models (2310.04041v2)

Published 6 Oct 2023 in cs.LG and cs.AI

Abstract: We propose a novel diffusion-based image generation method called the observation-guided diffusion probabilistic model (OGDM), which effectively addresses the tradeoff between quality control and fast sampling. Our approach reestablishes the training objective by integrating the guidance of the observation process with the Markov chain in a principled way. This is achieved by introducing an additional loss term derived from the observation based on a conditional discriminator on noise level, which employs a Bernoulli distribution indicating whether its input lies on the (noisy) real manifold or not. This strategy allows us to optimize the more accurate negative log-likelihood induced in the inference stage especially when the number of function evaluations is limited. The proposed training scheme is also advantageous even when incorporated only into the fine-tuning process, and it is compatible with various fast inference strategies since our method yields better denoising networks using the exactly the same inference procedure without incurring extra computational cost. We demonstrate the effectiveness of our training algorithm using diverse inference techniques on strong diffusion model baselines. Our implementation is available at https://github.com/Junoh-Kang/OGDM_edm.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (38)
  1. Estimating the optimal covariance with imperfect mean in diffusion probabilistic models. In ICML, 2022a.
  2. Analytic-DPM: An analytic estimate of the optimal reverse variance in diffusion probabilistic models. In ICLR, 2022b.
  3. Diffusion models beat GANs on image synthesis. In NeurIPS, 2021.
  4. GENIE: Higher-order denoising diffusion solvers. In NeurIPS, 2022.
  5. Generative adversarial nets. In NeurIPS, 2014.
  6. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In NeurIPS, 2017.
  7. Denoising diffusion probabilistic models. In NeurIPS, 2020.
  8. Equivariant diffusion for molecule generation in 3D. In ICML, 2022.
  9. Diff-TTS: A denoising diffusion model for text-to-speech. In INTERSPEECH, 2021.
  10. Gotta go fast when generating data with score-based models. arXiv preprint arXiv:2105.14080, 2021.
  11. Training generative adversarial networks with limited data. In NeurIPS, 2020.
  12. Elucidating the design space of diffusion-based generative models. In NeurIPS, 2022.
  13. Refining generative process with discriminator guidance in score-based diffusion models. In ICML, 2023.
  14. On fast sampling of diffusion probabilistic models. In ICML Workshop on Invertible Neural Networks, Normalizing Flows, and Explicit Likelihood Models, 2021.
  15. DiffWave: A versatile diffusion model for audio synthesis. In ICLR, 2021.
  16. Learning multiple layers of features from tiny images. 2009.
  17. Improved precision and recall metric for assessing generative models. 2019.
  18. Pseudo numerical methods for diffusion models on manifolds. In ICLR, 2022.
  19. Deep learning face attributes in the wild. In ICCV, 2015.
  20. Knowledge distillation in iterative generative models for improved sampling speed. arXiv preprint arXiv:2101.02388, 2021.
  21. Diffusion probabilistic models for 3D point cloud generation. In CVPR, 2021.
  22. On distillation of guided diffusion models. In CVPR, 2023.
  23. Improved denoising diffusion probabilistic models. In ICML, 2021.
  24. Permutation invariant graph generation via score-based generative modeling. In AISTATS, 2020.
  25. High-resolution image synthesis with latent diffusion models. In CVPR, 2022.
  26. Progressive distillation for fast sampling of diffusion models. In ICLR, 2022.
  27. Deep unsupervised learning using nonequilibrium thermodynamics. In ICML, 2015.
  28. Denoising diffusion implicit models. In ICLR, 2021a.
  29. Score-based generative modeling through stochastic differential equations. In ICLR, 2021b.
  30. Consistency models. In ICML, 2023.
  31. An Introduction to Numerical Analysis. Cambridge University Press, 1 edition, 2003.
  32. Rethinking the inception architecture for computer vision. In CVPR, 2016.
  33. Sauer Timothy. Numerical Analysis. Pearson, 3 edition, 2017.
  34. Diffusion-GAN: Training GANs with diffusion. 2023.
  35. Learning fast samplers for diffusion models by differentiating through sample quality. In ICLR, 2022.
  36. Tackling the generative learning trilemma with denoising diffusion GANs. In ICLR, 2022.
  37. LSUN: Construction of a large-scale image dataset using deep learning with humans in the loop. arXiv preprint arXiv:1506.03365, 2015.
  38. LION: Latent point diffusion models for 3D shape generation. In NeurIPS, 2022.

Summary

We haven't generated a summary for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com