Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 83 tok/s
Gemini 2.5 Pro 54 tok/s Pro
GPT-5 Medium 21 tok/s Pro
GPT-5 High 20 tok/s Pro
GPT-4o 103 tok/s Pro
Kimi K2 205 tok/s Pro
GPT OSS 120B 456 tok/s Pro
Claude Sonnet 4 35 tok/s Pro
2000 character limit reached

A Gradient Flow Approach to Solving Inverse Problems with Latent Diffusion Models (2509.19276v1)

Published 23 Sep 2025 in stat.ML, cs.LG, and stat.CO

Abstract: Solving ill-posed inverse problems requires powerful and flexible priors. We propose leveraging pretrained latent diffusion models for this task through a new training-free approach, termed Diffusion-regularized Wasserstein Gradient Flow (DWGF). Specifically, we formulate the posterior sampling problem as a regularized Wasserstein gradient flow of the Kullback-Leibler divergence in the latent space. We demonstrate the performance of our method on standard benchmarks using StableDiffusion (Rombach et al., 2022) as the prior.

Summary

  • The paper introduces DWGF, a training-free framework that leverages pretrained latent diffusion models to solve inverse problems using a gradient flow approach.
  • It formulates posterior sampling as a regularized Wasserstein gradient flow of the KL divergence in latent space, effectively balancing fidelity and diversity.
  • Experimental results show competitive PSNR and LPIPS metrics, though the method produces diverse yet blurry reconstructions with higher FID scores.

Gradient Flow Methods for Inverse Problems with Latent Diffusion Priors

Introduction

This paper introduces Diffusion-regularized Wasserstein Gradient Flow (DWGF), a training-free framework for solving ill-posed inverse problems by leveraging pretrained latent diffusion models (LDMs) as priors. The approach is motivated by the limitations of hand-crafted priors in high-dimensional settings and the computational inefficiency of pixel-space diffusion models. DWGF formulates posterior sampling as a regularized Wasserstein gradient flow (WGF) of the Kullback-Leibler (KL) divergence in the latent space, providing a principled method for Bayesian inference in inverse problems.

Methodology

Problem Formulation

Given an observation y=A(x0)+ϵy = \mathcal{A}(x_0) + \epsilon with known forward operator A\mathcal{A} and Gaussian noise ϵ\epsilon, the goal is to sample from the posterior p(x0y)p(yx0)pdata(x0)p(x_0|y) \propto p(y|x_0)p_{data}(x_0). In the context of LDMs, the prior is defined on a latent variable z0z_0, and the posterior is approximated as qμ(x0y)=pϕ(x0z0)μ(z0y)dz0q_\mu(x_0|y) = \int p_{\phi^-}(x_0|z_0)\mu(z_0|y)dz_0.

DWGF seeks the optimal latent posterior μ(z0y)\mu_\star(z_0|y) by minimizing a regularized functional:

μ(z0y)arg minμF[μ(z0y)]+γR[μ(z0y)]\mu_\star(z_0|y) \in \argmin_{\mu} \mathcal{F}[\mu(z_0|y)] + \gamma \mathcal{R}[\mu(z_0|y)]

where F\mathcal{F} is the KL divergence between the approximate and true posteriors, and R\mathcal{R} is a regularization term based on the diffusion prior.

Wasserstein Gradient Flow

The optimization is performed via WGF in the space of probability measures:

μtt=z0(μtz0δL[μ]δμ)\frac{\partial \mu_t}{\partial t} = \nabla_{z_0} \cdot \left(\mu_t \nabla_{z_0} \frac{\delta \mathcal{L}[\mu]}{\delta \mu}\right)

where L[μ]=F[μ]+γR[μ]\mathcal{L}[\mu] = \mathcal{F}[\mu] + \gamma \mathcal{R}[\mu]. The corresponding ODE for particles is:

dz0,tdt=z0δL[μ]δμ\frac{dz_{0,t}}{dt} = -\nabla_{z_0} \frac{\delta \mathcal{L}[\mu]}{\delta \mu}

Functional Derivatives

  • The gradient of F\mathcal{F} is derived using the reparameterization trick and chain rule, with the intractable data score x0logp(x0)\nabla_{x_0}\log p(x_0) approximated via the VAE encoder.
  • The regularization R\mathcal{R} is a weighted KL divergence along the diffusion process, with gradients computed via Monte Carlo over particles.

Algorithmic Implementation

  • The ODE is simulated using a particle-based approach, with gradients treated as loss terms and optimized using Adam for improved convergence.
  • Deterministic encoding is used due to negligible posterior variance, simplifying the data score computation.
  • The method is agnostic to the specific LDM architecture and can be extended to conditional settings.

Experimental Results

DWGF is evaluated on FFHQ-512 for box inpainting and 8×8\times super-resolution, compared against PSLD and RLSD baselines. Metrics include FID, PSNR, and LPIPS. Figure 1

Figure 1

Figure 1: Diversity of the particles x0x_0 produced by DWGF on large box inpainting.

DWGF achieves competitive PSNR and LPIPS but exhibits significantly higher FID, indicating less realistic sample quality. Qualitative analysis reveals that DWGF produces diverse but blurry reconstructions, attributed to the mode-seeking nature of the KL divergence. The particle cloud generated by DWGF demonstrates substantial diversity, but the lack of sharpness and realism is a notable limitation.

Implementation Considerations

  • Computational Requirements: The particle-based Monte Carlo approximation incurs substantial memory and compute overhead, especially for large NN and high-dimensional latent spaces.
  • Optimizer Choice: Adam is used for ODE discretization, introducing momentum and adaptive preconditioning, which is unconventional but empirically beneficial for ill-conditioned landscapes.
  • Regularization: The choice of regularization weight γ\gamma and data consistency parameter λ\lambda is critical for balancing fidelity and diversity.
  • Scalability: The method is compatible with few-step and consistency models, suggesting potential for acceleration and reduced sampling steps.

Limitations and Future Directions

DWGF does not match state-of-the-art FID scores, likely due to insufficient regularization and hyperparameter tuning. The framework is extensible to incorporate entropic regularization or repulsive potentials to mitigate mode collapse and improve sample realism. Integration with consistency models and control variates for variance reduction are promising avenues for future research. The particle-based approach may be further optimized for memory efficiency.

Theoretical Implications

DWGF provides a rigorous gradient flow perspective for posterior sampling in inverse problems, bridging optimal transport theory and generative modeling. The use of WGF in latent space enables principled regularization and efficient sampling, with potential applications in broader Bayesian inference tasks.

Conclusion

DWGF offers a principled, training-free approach for solving inverse problems with latent diffusion priors, leveraging Wasserstein gradient flows of KL divergence. While competitive in perceptual and reconstruction metrics, the method currently underperforms in sample realism. Future work should focus on enhanced regularization, integration with fast generative models, and improved computational efficiency. The gradient flow framework opens new directions for Bayesian inference and generative modeling in high-dimensional inverse problems.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 post and received 25 likes.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube