Improving Training-free Conditional Diffusion Model via Fisher Information (2404.18252v2)
Abstract: Training-free conditional diffusion models have received great attention in conditional image generation tasks. However, they require a computationally expensive conditional score estimator to let the intermediate results of each step in the reverse process toward the condition, which causes slow conditional generation. In this paper, we propose a novel Fisher information-based conditional diffusion (FICD) model to generate high-quality samples according to the condition. In particular, we further explore the conditional term from the perspective of Fisher information, where we show Fisher information can act as a weight to measure the informativeness of the condition in each generation step. According to this new perspective, we can control and gain more information along the conditional direction in the generation space. Thus, we propose the upper bound of the Fisher information to reformulate the conditional term, which increases the information gain and decreases the time cost. Experimental results also demonstrate that the proposed FICD can offer up to 2x speed-ups under the same sampling steps as most baselines. Meanwhile, FICD can improve the generation quality in various tasks compared to the baselines with a low computation cost.
- Denoising diffusion probabilistic models. CoRR, abs/2006.11239, 2020.
- Generative adversarial networks, 2014.
- Training-free content injection using h-space in diffusion models, 2024.
- Freedom: Training-free energy-guided conditional diffusion model. Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2023.
- Diffuse, attend, and segment: Unsupervised zero-shot segmentation using stable diffusion. CoRR, abs/2308.12469, 2023.
- Diffusion posterior sampling for general noisy inverse problems. In The Eleventh International Conference on Learning Representations, 2023.
- Adding conditional control to text-to-image diffusion models, 2023.
- Grad-cam: Visual explanations from deep networks via gradient-based localization. International Journal of Computer Vision, 128(2):336–359, October 2019.
- High-resolution image synthesis with latent diffusion models, 2021.
- Glide: Towards photorealistic image generation and editing with text-guided diffusion models, 2022.
- End-to-end diffusion latent optimization improves classifier guidance, 2023.
- Sdedit: Guided image synthesis and editing with stochastic differential equations, 2022.
- Plug-and-play diffusion features for text-driven image-to-image translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 1921–1930, June 2023.
- Pseudoinverse-guided diffusion models for inverse problems. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023, 2023.
- Manifold preserving guided diffusion, 2023.
- Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021.
- Denoising diffusion implicit models, 2022.
- Improving diffusion models for inverse problems using manifold constraints, 2022.
- Solving inverse problems with latent diffusion models via hard data consistency, 2023.
- Andrew R. Barron. Entropy and the central limit theorem. The Annals of Probability, 14(1):336–342, 1986.
- Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023, Vancouver, BC, Canada, June 17-24, 2023, pages 22500–22510, 2023.
- A latent space of stochastic diffusion models for zero-shot image editing and guidance. In ICCV, 2023.
- Adapt and diffuse: Sample-adaptive reconstruction via latent diffusion models, 2023.
- Information-theoretic diffusion. In International Conference on Learning Representations, 2023.
- InfoDiffusion: Representation learning using information maximizing diffusion models. In Proceedings of the 40th International Conference on Machine Learning, pages 36336–36354, 2023.
- Interpretable diffusion via information decomposition, 2023.
- Progressive growing of gans for improved quality, stability, and variation. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018.
- Learning transferable visual models from natural language supervision, 2021.