Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Prompt-tuning latent diffusion models for inverse problems (2310.01110v1)

Published 2 Oct 2023 in cs.LG, cs.AI, cs.CV, and stat.ML

Abstract: We propose a new method for solving imaging inverse problems using text-to-image latent diffusion models as general priors. Existing methods using latent diffusion models for inverse problems typically rely on simple null text prompts, which can lead to suboptimal performance. To address this limitation, we introduce a method for prompt tuning, which jointly optimizes the text embedding on-the-fly while running the reverse diffusion process. This allows us to generate images that are more faithful to the diffusion prior. In addition, we propose a method to keep the evolution of latent variables within the range space of the encoder, by projection. This helps to reduce image artifacts, a major problem when using latent diffusion models instead of pixel-based diffusion models. Our combined method, called P2L, outperforms both image- and latent-diffusion model-based inverse problem solvers on a variety of tasks, such as super-resolution, deblurring, and inpainting.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Hyungjin Chung (38 papers)
  2. Jong Chul Ye (210 papers)
  3. Peyman Milanfar (64 papers)
  4. Mauricio Delbracio (36 papers)
Citations (24)
X Twitter Logo Streamline Icon: https://streamlinehq.com