Diffusion Priors in Inverse Problems
- Diffusion priors are SDE-parameterized probabilistic models that learn score functions to capture complex geometric, semantic, and multimodal features.
- They integrate data-driven priors with Bayesian inverse problem formulations, using reverse SDE sampling and likelihood gradients for robust reconstruction.
- They excel in handling sparse, noisy, or partial data, yielding higher fidelity outcomes compared to conventional hand-crafted priors.
A diffusion prior is a probabilistic model, parameterized via a diffusion process (forward SDE and learned reverse SDE), used to inject powerful, data-driven structural constraints into generative and inverse problems. Unlike traditional priors (e.g., sparsity, smoothness), diffusion priors are learned from data and are capable of capturing the intricate geometric, semantic, or multimodal structure of complex domains such as images, 3D objects, or biological assemblies.
1. Diffusion Priors: Formalism and Scoring
A diffusion prior is constructed by training a stochastic differential equation (SDE)-based generative model. The forward SDE corrupts samples into isotropic noise over time, typically: In the Karras–EDM formulation, and ; so diffuses to Gaussian noise as .
The reverse SDE (Anderson 1982) moves from pure noise back to data: where is the marginal at diffusion time , and is the time-dependent Stein score. The score function is estimated by training a neural network via denoising score matching (DSM), using Gaussian perturbation and minimizing
with the analytical score (Möbius et al., 2024).
2. Diffusion Priors in Bayesian Inverse Problems
Given incomplete or noisy observations arising from an unknown through
the Bayesian posterior is
The diffusion prior encodes the data manifold more expressively than classical hand-crafted priors.
To incorporate the likelihood, one defines an energy for the forward operator and writes
allowing for flexible data terms (e.g., permutation-invariant matching of projections). Posterior sampling leverages a reverse SDE augmented by likelihood gradients: where is the denoised guess, and balances prior and data fit. The resulting "Diffusion Posterior Sampling" iteratively applies Euler–Maruyama or other discretizations that couple score and data likelihood (Möbius et al., 2024).
3. Implementation Details and Posterior Sampling
The general sampling algorithm proceeds as follows:
- Initialization: Sample .
- Reverse SDE Step: For discrete times , with ,
- Compute prior drift using the approximate posterior score
- Apply a second-order correction and inject noise
- Repeat until
Concretely,
with and , as in the main text (Möbius et al., 2024).
Data terms are application-specific but always enter as gradients in the posterior score.
4. Application Domains and Empirical Advantages
Diffusion priors have been applied in 3D reconstruction from incomplete projections (e.g., cryo-EM), image restoration, medical imaging, and other ill-posed inverse problems:
- Cryo-EM/ShapeNet: Reconstruction from as few as 1–5 projections and/or low-resolution traces using diffusion priors achieves RMSD reductions from 5–12 Å (ML) to 3–8 Å (DPS), and halves error metrics such as Chamfer Distance and EMD relative to classical maximum-likelihood (Möbius et al., 2024).
- Resolution Regimes: Diffusion priors are especially beneficial with very sparse, noisy, or partial data, enabling "intermediate-resolution" reconstructions that are unattainable with hand-crafted priors.
The table below exemplifies the gains in molecular assembly reconstruction:
| Regime | ML RMSD (Å) | DPS RMSD (Å) |
|---|---|---|
| Single view + coarse points | 5–12 | 3–8 |
| 4–6 projections (ShapeNet chairs) | High error | ~50% reduction |
| Cryo-EM (Proteins) | 5–12 | 3–8 |
The unified reverse SDE sampling brings together generative priors and experimental constraints, yielding improved sample quality and data fidelity, especially in strongly underdetermined settings (Möbius et al., 2024).
5. Generalization, Limitations, and Future Directions
Diffusion priors exhibit:
- Generalization: Flexible adaptation to varied forward models and noise regimes via redefinition of the likelihood and score coupling.
- Complex Structure Capture: Unlike simple priors, diffusion models fit non-trivial manifolds such as biologically realistic 3D forms or textile texture.
- No closed-form likelihood: The prior is only accessible through the learned score; direct evaluation of is intractable—one must sample.
- Sampling overhead: Posterior sampling is considerably more expensive than with analytic priors, though methods such as second-order or flow-based acceleration are under active development.
Open questions include optimal score interpolation, balancing prior vs. likelihood (schedule ), and rigorous convergence analysis of the posterior sampler.
6. Theoretical and Methodological Foundations
Diffusion priors provide a data-driven density defined through the integrated reverse-time SDE, with score-learning rooted in denoising score matching:
- Posterior formulation:
- Training criterion:
- Reverse SDE:
- Posterior score:
Empirically, these constructs harmonize denoiser priors with experimental likelihoods such that solutions interpolate flexibly between strong generative prior support and observed data (Möbius et al., 2024).
In summary, a diffusion prior is a data-derived, SDE-parametrized probabilistic model leveraged as the prior in a Bayesian inverse problem framework. By learning a rich score and unifying with the forward-model likelihood in a reverse SDE-driven sampler, diffusion priors enable high-fidelity solutions for ill-posed problems inaccessible to conventional priors—especially when observations are highly incomplete, noisy, or partial (Möbius et al., 2024).