Papers
Topics
Authors
Recent
2000 character limit reached

Latent Diffusion-Based Differentiable Inversion

Updated 3 January 2026
  • The paper introduces LD-DIM, which couples pretrained latent diffusion priors with fully differentiable PDE solvers to robustly reconstruct high-dimensional parameter fields.
  • It leverages an encoder-decoder architecture and adjoint-based gradient computation to optimize in a low-dimensional latent space, reducing ill-conditioning.
  • Empirical evaluations demonstrate that LD-DIM outperforms PINNs and VAE-based methods, achieving up to 30× error reduction and superior SSIM scores.

The latent diffusion-based differentiable inversion method (LD-DIM) is a framework for solving high-dimensional inverse problems, particularly those governed by partial differential equations (PDEs), by coupling pretrained latent diffusion priors with fully differentiable numerical solvers. LD-DIM achieves stable and accurate reconstruction of spatially heterogeneous parameter fields from sparse observations through gradient-based optimization performed directly in the low-dimensional latent space of a diffusion model. Incorporating adjoint-based gradient computation, end-to-end automatic differentiation, and physics-based solvers, LD-DIM achieves significant improvements in conditioning, accuracy, and robustness, outperforming prevalent alternatives such as physics-informed neural networks (PINNs) and physics-embedded variational autoencoders (VAEs) for diverse inverse modeling tasks (Lin et al., 27 Dec 2025, Rout et al., 2023, Wang et al., 2024, Bai et al., 2024).

1. Latent Diffusion Priors and Model Architecture

LD-DIM relies on a latent diffusion model (LDM) that models the distribution of high-dimensional parameter fields in a low-dimensional latent space. The primary components are:

  • Encoder EϕE_\phi: A convolutional variational autoencoder (VAE) encoder maps the high-dimensional parameter field (e.g., conductivity K(x)RH×WK(\mathbf{x}) \in \mathbb{R}^{H \times W}) to a latent Gaussian posterior qϕ(z0K)q_\phi(z_0|K), with z0Rdz_0 \in \mathbb{R}^d.
  • Decoder DψD_\psi: A convolutional decoder reconstructs the field, K^=Dψ(z0)\hat{K} = D_\psi(z_0), ensuring that all fields lie on the learned manifold {Dψ(z):zRd}\{ D_\psi(z) : z \in \mathbb{R}^d \}.
  • Score Network εθ(zt,t)\varepsilon_\theta(z_t, t): A U-Net in latent space, trained for denoising score matching to approximate ztlogpt(zt)\nabla_{z_t} \log p_t(z_t) at each timestep tt.

The training objective comprises a VAE loss for reconstruction and posterior regularization,

LVAE=KDψ(z0)1+λKLDKL(qϕ(z0K)    N(0,I)),\mathcal{L}_{\mathrm{VAE}} = \| K - D_\psi(z_0) \|_{1} + \lambda_{\mathrm{KL}} D_{\mathrm{KL}}(q_\phi(z_0|K)\;||\;\mathcal{N}(0,I)),

and a diffusion loss,

Ldiff=Et,z0,εεεθ(αˉtz0+1αˉtε,t)22,\mathcal{L}_{\rm diff} = \mathbb{E}_{t,z_0,\varepsilon} \| \varepsilon - \varepsilon_\theta(\sqrt{\bar\alpha_t} z_0 + \sqrt{1-\bar\alpha_t} \varepsilon, t) \|^2_2,

combined to define the variational bound on the data log-likelihood,

L=LVAE+Ldifflogp(K).\mathcal{L} = \mathcal{L}_{\mathrm{VAE}} + \mathcal{L}_{\mathrm{diff}} \leq -\log p(K).

This latent prior captures complex variability and, crucially, preserves interfaces and sharp features, addressing the sharpness deficit of pixel-based generative approaches (Lin et al., 27 Dec 2025, Rout et al., 2023).

2. Differentiable PDE Solvers and Adjoint Techniques

Inverse modeling in LD-DIM is performed by coupling the LDM with a fully differentiable forward solver for the governing PDE. For example, in subsurface flow scenarios:

  • Governing equation:

(K(x)h(x))=0\nabla \cdot (K(\mathbf{x}) \nabla h(\mathbf{x})) = 0

(Darcy flow).

  • Numerical discretization: A finite-volume (FVM) scheme with two-point flux approximation results in a sparse linear system A(K)h=bA(K)h = b where A(K)A(K) is assembled and solved using efficient sparse solvers (e.g. in JAX).
  • Adjoint state computation: Sensitivities are backpropagated via the discrete adjoint system

A(K)λ=r,A(K)^\top \lambda = r,

allowing efficient construction of gradients for objectives involving both PDE misfit and diffusion priors (Lin et al., 27 Dec 2025, Wang et al., 2024).

Automatic differentiation frameworks (JAX) are employed for propagating gradients through the decoder DψD_\psi and the PDE solver, with custom vector-Jacobian product (VJP) implementations to avoid storage of large dense Jacobians and reuse the linear system in the reverse pass.

3. Inversion and Optimization in the Latent Space

The entire inversion procedure operates in latent space, drastically reducing optimization dimensionality and conditioning pathologies common in pixel or coefficient space:

  • Objective function: For data hh^\star observed at indices Iobs\mathcal{I}_{\rm obs},

J(z)=(z)+βR(z),(z)=iIobs(hihi(z))2,R(z)=12z22J(z) = \ell(z) + \beta R(z), \quad \ell(z) = \sum_{i \in \mathcal{I}_{\rm obs}} (h^\star_i - h_i(z))^2, \quad R(z) = \frac{1}{2}\|z\|_2^2

with hi(z)h_i(z) denoting the predicted PDE solution decoded from latent zz.

  • Gradient computation: Gradients with respect to latent zz are computed efficiently using the chain rule through the decoder and solver; adjoint methods further improve computational efficiency.
  • Optimization: Latent codes zz are iteratively updated via Adam or SGD; typical convergence is obtained in a few hundred steps, starting from z(0)N(0,I)z^{(0)} \sim \mathcal{N}(0,I) (Lin et al., 27 Dec 2025).

By restricting search to the non-linear manifold of plausible fields learned by the LDM, the method implicitly regularizes the problem, suppresses ill-conditioned modes, and enables the accurate resolution of dominant and sharp structures.

4. Comparative Performance and Empirical Evaluation

LD-DIM has been empirically evaluated on representative PDE-constrained inverse problems, including:

  • Random field and bimaterial subsurface conductivity reconstruction: Outperforms PINNs and VAE-based methods in both L2L^2 error and the structural similarity index (SSIM), e.g., reducing ϵK\epsilon_K by up to 30×30\times and raising SSIM from near zero to over 0.9 compared to PINNs.
  • Sensitivity and seed robustness: Multiple optimization runs from different zz seeds yield consistent reconstructions with only minor shifts at sharp interfaces, demonstrating robustness to initialization.
  • Observation density: As the number of observed data points increases (from 3×33\times3 to 16×1616\times16 in grid settings), reconstruction accuracy and variance improve, but even with very sparse observations LD-DIM captures large-scale structure (Lin et al., 27 Dec 2025).

Similarly, for canonical linear inverse tasks (inpainting, super-resolution), latent-space LD-DIM posterior sampling outperforms pixel-space diffusion approaches (DPS, DDRM) across PSNR, SSIM, and LPIPS (Rout et al., 2023).

5. Extensions and Applications in Physics-Constrained and Blind Inverse Problems

LD-DIM generalizes to various modalities and physical constraints:

  • Joint generative latent spaces for multimodal PDE problems: In full waveform inversion (FWI), a shared latent codebook is used for both seismic and velocity fields, with the diffusion process enforcing consistency with the governing PDE. The learned prior acts as an implicit physics regularizer, and denoising steps refine solutions toward physical feasibility (Wang et al., 2024).
  • Blind inverse problems (unknown forward operator): LD-DIM is extended to alternating EM frameworks. The E-step samples from the latent posterior using prior and likelihood gradients; the M-step updates forward operator parameters. Annealing and skip-gradient acceleration techniques improve robustness and efficiency in the EM loop (Bai et al., 2024).

LD-DIM, thus, unifies Bayesian inference with deep generative priors and physics-based simulation, enabling direct integration of uncertainty quantification, Bayesian posterior sampling, and full probabilistic inversion, as well as scalability to 3D and time-dependent problems.

6. Algorithmic Workflow and Computational Considerations

A pseudocode outlining the LD-DIM workflow is as follows (Lin et al., 27 Dec 2025, Rout et al., 2023):

Step Operation Purpose
1. Train VAE Pretrain {Eϕ,Dψ}\{E_\phi, D_\psi\} on field dataset via LVAE\mathcal{L}_{\mathrm{VAE}} Learn low-dimensional manifold
2. Train LDM Fix VAE, train latent-space U-Net εθ\varepsilon_\theta with Ldiff\mathcal{L}_{\mathrm{diff}} Model prior over plausible fields
3. Initialize zz z(0)N(0,I)z^{(0)} \sim \mathcal{N}(0, I) Random initialization
4. Optimization loop Decode K=Dψ(z)K = D_\psi(z); Solve A(K)h=bA(K)h = b; Compute \ell, JJ; Compute adjoint λ\lambda; Update zz Minimize loss in latent space
5. Final reconstruction Output K^=Dψ(z)\hat{K} = D_\psi(z^\star) Recovered field

Computationally, VAE and diffusion model pretraining requires moderate GPU time (hours for typical datasets); each inversion involves O(102)O(10^2) gradient steps, with per-step cost dominated by the sparse PDE solve and decoder pass. The method is inherently scalable and compatible with modern autodiff frameworks (Lin et al., 27 Dec 2025).

7. Limitations, Practical Issues, and Future Directions

While LD-DIM exhibits robust numerical conditioning and empirical performance, certain limitations persist:

  • Data sparsity: Large-scale structure is reconstructable from minimal measurements, but recovery of fine-scale features necessitates higher observation density (>5×5>5 \times 5 grid points) (Lin et al., 27 Dec 2025).
  • Computational overhead: Cost arises mainly from the PDE solver and diffusion prior evaluations, but is manageable for moderate NN.
  • Extensions: Natural generalizations include application to time-dependent and multiphase PDEs, fully 3D domains, and integration with explicit noise models or Bayesian sampling in the latent posterior (Lin et al., 27 Dec 2025, Wang et al., 2024, Bai et al., 2024).

Recent studies suggest that latent diffusion priors not only regularize ill-posed inverse problems but, when coupled with proper physical constraints, recover solutions that preserve fine structure and physical realism, with credible uncertainty quantification (Rout et al., 2023, Wang et al., 2024). The framework offers a principled and scalable toolset for high-dimensional scientific inversion under data scarcity and modeling uncertainty.

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Latent Diffusion-Based Differentiable Inversion Method (LD-DIM).