Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Manifold Preserving Guided Diffusion (2311.16424v1)

Published 28 Nov 2023 in cs.LG, cs.AI, and cs.CV

Abstract: Despite the recent advancements, conditional image generation still faces challenges of cost, generalizability, and the need for task-specific training. In this paper, we propose Manifold Preserving Guided Diffusion (MPGD), a training-free conditional generation framework that leverages pretrained diffusion models and off-the-shelf neural networks with minimal additional inference cost for a broad range of tasks. Specifically, we leverage the manifold hypothesis to refine the guided diffusion steps and introduce a shortcut algorithm in the process. We then propose two methods for on-manifold training-free guidance using pre-trained autoencoders and demonstrate that our shortcut inherently preserves the manifolds when applied to latent diffusion models. Our experiments show that MPGD is efficient and effective for solving a variety of conditional generation applications in low-compute settings, and can consistently offer up to 3.8x speed-ups with the same number of diffusion steps while maintaining high sample quality compared to the baselines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Yutong He (43 papers)
  2. Naoki Murata (29 papers)
  3. Chieh-Hsin Lai (32 papers)
  4. Yuhta Takida (32 papers)
  5. Toshimitsu Uesaka (17 papers)
  6. Dongjun Kim (24 papers)
  7. Wei-Hsiang Liao (33 papers)
  8. Yuki Mitsufuji (127 papers)
  9. J. Zico Kolter (151 papers)
  10. Ruslan Salakhutdinov (248 papers)
  11. Stefano Ermon (279 papers)
Citations (35)

Summary

We haven't generated a summary for this paper yet.