SURE Guided Posterior Sampling (SGPS)
- SGPS is a trajectory-corrected inference algorithm that integrates diffusion denoising with SURE-based error correction and PCA noise estimation.
- It interleaves conditional posterior guidance with local residual measurement to achieve high-quality image reconstructions under tight computational budgets.
- The method leverages unbiased risk estimates and KL convergence theory to mitigate error accumulation, ensuring efficient correction during sampling.
SURE Guided Posterior Sampling (SGPS) is a trajectory-corrected inference algorithm for diffusion-based inverse problems that leverages Stein’s Unbiased Risk Estimate (SURE) and PCA-based noise estimation to mitigate error accumulation in the critical early and middle stages of sampling. SGPS consistently achieves high-quality reconstructions under tight computational budgets—requiring fewer than 100 Neural Function Evaluations (NFEs)—by interleaving diffusion denoising, conditional posterior guidance, local residual noise measurement, and data-guided correction steps (Kim et al., 29 Dec 2025).
1. Inverse Problem Formulation and Diffusion Priors
The core objective is to recover an unknown image from noisy linear measurements: where is a known forward operator (e.g., for super-resolution or deblurring), and is the measurement noise standard deviation.
A diffusion model serves as the learned prior, specified by a stochastic differential equation (SDE): with reparameterization in EDM by .
The unconditional backward sampling (reverse-time ODE) is: approximated in practice via a pre-trained denoiser :
Posterior sampling for the inverse problem uses Bayes’ rule: The data-consistency term is typically intractable and must be approximated.
2. SURE-Based Trajectory Correction
2.1 Stein's Unbiased Risk Estimate (SURE)
SURE provides an unbiased estimator of the mean squared error (MSE) for denoising under additive Gaussian noise. For , , and denoiser : where . The trace is estimated by a Monte Carlo probe:
The SURE gradient direction is obtained by differentiating SURE w.r.t. : The correction is: where is a user-chosen step size; experiments use .
2.2 Local SURE Gradient Update
After applying conditional guidance, let be the resulting state. Given a residual noise estimate (see Section 3), the denoiser is applied: with SURE evaluated as: A correction step via autodiff follows, reducing residual noise and pulling samples toward the data manifold.
3. PCA-Based Residual Noise Estimation
Accurate SURE application requires knowledge of the residual variance in . SGPS employs a patch PCA estimator:
- Decompose into overlapping patches , compute mean and covariance
- Eigen-decompose to obtain eigenvalues . For each , define: The smallest with equal to the median of is chosen. The noise level is then
This estimator is efficient and requires no additional training.
4. SURE Guided Posterior Sampling Algorithm
The SGPS algorithm proceeds as follows:
- Initialization: Sample .
- For :
- a) Denoising: .
- b) Conditional Guidance: Use Langevin iterations to obtain that balances prior and data likelihood.
- c) PCA Noise Estimation: Estimate residual noise from .
- d) SURE Gradient Correction: Apply local correction using the SURE gradient to , yielding .
- e) Sample for Next Step: .
- Return .
Distinctive features:
- Estimated, not assumed, noise levels at each step ( via PCA).
- Local SURE-based correction at every iteration directly addresses sampling trajectory deviations.
5. Theoretical Properties
- Gaussian-Preservation (Theorem 1): Small-step Langevin guidance ensures the output of the denoiser remains nearly Gaussian in Wasserstein-2 distance , justifying the use of SURE at each iteration.
- KL-Convergence with SURE Correction (Theorem 2): Under local strong convexity of and bounded SURE bias/variance, each correction step reduces the KL divergence to the true posterior, up to error, where :
- Error-Cascade Mitigation: By removing residual noise at each iteration, SGPS avoids error accumulation characteristic of earlier-stage high-noise samples, enabling accurate inference with NFEs.
6. Empirical Performance and Cost Analysis
6.1 Benchmark Domains
SGPS was evaluated on linear (FFHQ256 super-resolution , box inpainting, random inpainting, Gaussian and motion deblurring) and nonlinear (phase retrieval, nonlinear deblurring, HDR recovery) inverse problems.
6.2 Quantitative Results
Performance with ( NFE) and ( NFE) is reported using PSNR (higher is better) and LPIPS (lower is better):
| Method | NFE | SR4 PSNR / LPIPS | InpaintBox PSNR / LPIPS | InpaintRnd PSNR / LPIPS | GaussDebl PSNR / LPIPS | MotDebl PSNR / LPIPS |
|---|---|---|---|---|---|---|
| SGPS | 99 | 29.38 / 0.179 | 24.23 / 0.133 | 30.47 / 0.116 | 29.35 / 0.179 | 31.24 / 0.148 |
| DAPS | 100 | 27.69 / 0.230 | 22.51 / 0.192 | 26.64 / 0.238 | 27.77 / 0.220 | 29.84 / 0.167 |
| Method | NFE | PhaseRet PSNR / LPIPS | NonlinDebl PSNR / LPIPS | HDR PSNR / LPIPS |
|---|---|---|---|---|
| SGPS | 99 | 24.08 / 0.268 | 27.33 / 0.197 | 24.87 / 0.179 |
| DAPS | 100 | 20.83 / 0.402 | 25.56 / 0.255 | 24.09 / 0.199 |
6.3 Computational Cost
- On an RTX 4090: 48 NFE 4.13 s/image, 99 NFE 8.46 s/image.
- In competitive SR4 settings at comparable runtime (4 s), SGPS achieves PSNR 29.06 dB versus DDNM's 29.09 dB.
- Overhead breakdown (for 48 NFE): SURE update (denoiser + autograd) 51.2%, Langevin guidance 35.5%, forward denoise 11.2%, PCA 1.8%.
7. Implementation Considerations and Limitations
- Denoiser: U-Net in VP-DDPM/EDM configuration, trained on FFHQ256 images.
- Noise schedule: Geometric, from to , with (Karras et al.).
- Sampling Steps: (48 NFE), (99 NFE).
- Langevin Conditional Guidance: 100 iterations per outer step, step size .
- PCA: Patch size , stride 4, patches per image.
- SURE Hyperparameters: , .
- Trace Vectors: One random vector per step; additional vectors confer no empirical benefit.
Principal limitations include restriction to pixel-space diffusion samplers, an assumption of known (forward operator), and requirement of local strong convexity for convergence theory. PCA noise estimation may fail for images with little self-similarity; alternative estimators (e.g., spectral) are potential directions. The SURE update uses backpropagation; forward-mode JVP or SPSA could reduce cost. Blind or partially unknown forward operators, non-Gaussian noise, and adaptation to latent-diffusion models remain open areas.
For detailed derivations and algorithmic implementations, see (Kim et al., 29 Dec 2025).