Improving Diffusion-Based Generative Models via Approximated Optimal Transport (2403.05069v1)
Abstract: We introduce the Approximated Optimal Transport (AOT) technique, a novel training scheme for diffusion-based generative models. Our approach aims to approximate and integrate optimal transport into the training process, significantly enhancing the ability of diffusion models to estimate the denoiser outputs accurately. This improvement leads to ODE trajectories of diffusion models with lower curvature and reduced truncation errors during sampling. We achieve superior image quality and reduced sampling steps by employing AOT in training. Specifically, we achieve FID scores of 1.88 with just 27 NFEs and 1.73 with 29 NFEs in unconditional and conditional generations, respectively. Furthermore, when applying AOT to train the discriminator for guidance, we establish new state-of-the-art FID scores of 1.68 and 1.58 for unconditional and conditional generations, respectively, each with 29 NFEs. This outcome demonstrates the effectiveness of AOT in enhancing the performance of diffusion models.
- Wasserstein generative adversarial networks. International Conference on Machine Learning, 2017.
- Computer methods for ordinary differential equations and differential-algebraic equations, volume 61. Siam, 1998.
- Generative adversarial networks. Communications of the ACM, 63(11):139–144, 2020.
- Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017.
- Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840–6851, 2020.
- Exponential integrators. Acta Numerica, 19:209–286, 2010.
- A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4401–4410, 2019.
- Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 8110–8119, 2020.
- Elucidating the design space of diffusion-based generative models. Advances in Neural Information Processing Systems, 35:26565–26577, 2022.
- Understanding DDPM latent codes through optimal transport. In The Eleventh International Conference on Learning Representations, 2023.
- Refining generative process with discriminator guidance in score-based diffusion models. International Conference on Machine Learning, 2023.
- Consistency trajectory models: Learning probability flow ode trajectory of diffusion. arXiv preprint arXiv:2310.02279, 2023.
- Learning multiple layers of features from tiny images. 2009.
- Harold W Kuhn. The hungarian method for the assignment problem. Naval research logistics quarterly, 2(1-2):83–97, 1955.
- Flow matching for generative modeling. In International Conference on Learning Representations, 2023.
- Pseudo numerical methods for diffusion models on manifolds. In International Conference on Learning Representations, 2022.
- Flow straight and fast: Learning to generate and transfer data with rectified flow. In The Eleventh International Conference on Learning Representations, 2023.
- Multisample flow matching: Straightening flows with minibatch couplings. International Conference on Machine Learning, 2023.
- Hierarchical text-conditional image generation with clip latents, 2022.
- High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 10684–10695, 2022.
- U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pages 234–241. Springer, 2015.
- Photorealistic text-to-image diffusion models with deep language understanding. Advances in Neural Information Processing Systems, 35:36479–36494, 2022.
- Timothy Sauer. Numerical solution of stochastic differential equations in finance. In Handbook of computational finance, pages 529–550. Springer, 2011.
- Generative modeling by estimating gradients of the data distribution. Advances in neural information processing systems, 32, 2019.
- Denoising diffusion implicit models. In International Conference on Learning Representations, 2020.
- Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021.
- Consistency models. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett, editors, Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pages 32211–32252. PMLR, 23–29 Jul 2023.
- Conditional flow matching: Simulation-free dynamic optimal transport. arXiv preprint arXiv:2302.00482, 2023.
- Score-based generative modeling in latent space. In Neural Information Processing Systems (NeurIPS), 2021.
- Cédric Villani et al. Optimal transport: old and new, volume 338. Springer, 2009.
- Cedric Villani. Topics in optimal transportation.(books). OR/MS Today, 30(3):66–67, 2003.
- Poisson flow generative models. Advances in Neural Information Processing Systems, 35:16782–16795, 2022.
- Pfgm++: Unlocking the potential of physics-inspired generative models. International Conference on Machine Learning, 2023.
- Stable target field for reduced variance score estimation in diffusion models. In International Conference on Learning Representations, 2023.
- Fast sampling of diffusion models with exponential integrator. In International Conference on Learning Representations, 2023.
- Contrastive sampling chains in diffusion models. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.