Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Dimensionality-Varying Diffusion Process (2211.16032v1)

Published 29 Nov 2022 in cs.LG and cs.CV

Abstract: Diffusion models, which learn to reverse a signal destruction process to generate new data, typically require the signal at each step to have the same dimension. We argue that, considering the spatial redundancy in image signals, there is no need to maintain a high dimensionality in the evolution process, especially in the early generation phase. To this end, we make a theoretical generalization of the forward diffusion process via signal decomposition. Concretely, we manage to decompose an image into multiple orthogonal components and control the attenuation of each component when perturbing the image. That way, along with the noise strength increasing, we are able to diminish those inconsequential components and thus use a lower-dimensional signal to represent the source, barely losing information. Such a reformulation allows to vary dimensions in both training and inference of diffusion models. Extensive experiments on a range of datasets suggest that our approach substantially reduces the computational cost and achieves on-par or even better synthesis performance compared to baseline methods. We also show that our strategy facilitates high-resolution image synthesis and improves FID of diffusion model trained on FFHQ at $1024\times1024$ resolution from 52.40 to 10.46. Code and models will be made publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Han Zhang (338 papers)
  2. Ruili Feng (21 papers)
  3. Zhantao Yang (8 papers)
  4. Lianghua Huang (19 papers)
  5. Yu Liu (786 papers)
  6. Yifei Zhang (167 papers)
  7. Yujun Shen (111 papers)
  8. Deli Zhao (66 papers)
  9. Jingren Zhou (198 papers)
  10. Fan Cheng (19 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.