GABI: Geometric Autoencoders for Bayesian Inversion
- The paper introduces a two-stage 'learn first, observe later' framework that uses geometry-aware autoencoders to capture latent representations of physical fields.
- It achieves scalable Bayesian inversion by mapping high-dimensional problems to low-dimensional latent spaces with geometry-conditioned generative priors.
- Significant applications in heat transfer, airfoil flows, and 3D resonance demonstrate competitive accuracy and robust uncertainty quantification.
Geometric Autoencoders for Bayesian Inversion (GABI) constitute a framework for uncertainty quantification and inference in high-dimensional inverse problems, specifically targeting scenarios with variable and complex geometries. The method leverages geometry-aware autoencoder architectures to parameterize the space of possible physical responses and then fuses this geometry-adapted prior with a likelihood informed by noisy, sparse observations to obtain Bayesian posterior estimates. GABI enacts a “learn first, observe later” paradigm, separating large-scale unsupervised model learning from downstream adaptation to specific inference tasks (Vadeboncoeur et al., 24 Sep 2025).
1. Framework Overview
GABI is formulated as a two-stage process exploiting geometric deep learning and generative modeling:
- Learning Stage (“learn first”): A geometry-conditioned autoencoder is trained on a dataset of full-field responses—each associated with distinct system geometries, boundary conditions, and possibly different physical parameters. The encoder maps a physical field (e.g., temperature, pressure, vibration amplitude) defined on discretized geometry to a latent vector . The decoder inverts this mapping, reconstructing the original field on the same geometry.
- Inference Stage (“observe later”): Upon receiving a new geometry and observations modeled as , with and noise , the autoencoder’s latent prior is combined with the likelihood of the observed data in the Bayesian framework. The key result (Lemma 2.1) is that, under the pushforward prior $u \sim D_o^\psi_{\#} q_z$, the posterior for is the pushforward of the latent space posterior: $p(u_o|y_o) = D_o^\psi_{\#} p(z|y_o)$.
This structure enables the training of a foundation model that is fully geometry-aware and independent of the specific observation process, supporting broad generalization across observation types and spatial domains (Vadeboncoeur et al., 24 Sep 2025).
2. Geometry-Conditioned Generative Priors
The central technical innovation in GABI is the use of autoencoders whose encoder and decoder architectures explicitly respect and operate on variable mesh, graph, or discretization data structures representing the underlying physical geometries. The encoder and decoder are parameterized to accept geometric context as input, ensuring that latent variables encode information relevant to the geometry . The prior on each field is formulated as the pushforward of a simple latent distribution (typically ) through the geometry-conditioned decoder:
This design trains the latent prior to capture not only low-dimensional structures of the response fields but also regularities and dependencies induced by the geometric variability of the training data. Notably, the autoencoder is optimized by minimizing:
where is a chosen divergence (e.g., maximum mean discrepancy) enforcing alignment between the empirical distribution of latent codes and the fixed prior (Vadeboncoeur et al., 24 Sep 2025).
3. Bayesian Inversion in the Latent Space
Once trained, GABI enables Bayesian inversion by translating inference from the ambient field space to the fixed, low-dimensional latent space. Given observations on a new geometry :
- The observation model is expressed as , leading to a likelihood .
- The latent posterior is computed as .
- Posterior samples from are propagated through the decoder to yield posterior field samples .
This approach permits efficient sampling-based inference in (often is on the order of $10$–$100$) as opposed to the much higher-dimensional field spaces, rendering the inversion computationally tractable even in the presence of complex geometric context and high observational noise (Vadeboncoeur et al., 24 Sep 2025).
4. Implementation: Sampling and Scalability
GABI employs both Markov Chain Monte Carlo (using the No-U-Turn Sampler) and Approximate Bayesian Computation (ABC) as posterior sampling methodologies in the latent space. ABC, in particular, capitalizes on extensive GPU acceleration by proposing a large ensemble of latent codes, decoding them in parallel, and selecting samples whose predicted observations are within a user-specified residual threshold. This leverages modern computational hardware, enabling practical uncertainty quantification at scale for large engineering domains and variable geometries.
GABI’s encoder-decoder implementations draw on geometric deep learning: graph convolutional networks and transformer blocks are commonly used to process non-Euclidean discretizations, supporting variable mesh sizes and allowing for flexible mesh input/output shapes (Vadeboncoeur et al., 24 Sep 2025).
5. Applications and Empirical Validation
The framework has been validated on multiple large-scale engineering and physics problems characterized by both ill-posedness and substantial geometric variabilities:
- Steady-State Heat: GABI is trained over $1,000$ distinct rectangular domains with varying boundary conditions. Posterior predictive fields on new domains, inferred from a handful of noisy measurements, achieve accuracy competitive with supervised regression baselines, while uncertainty quantification is robust and well calibrated.
- RANS Flow Around Airfoils: Using graph-based autoencoder architectures, GABI reconstructs the full pressure/velocity fields on airfoils of previously unseen shape, conditioned on sparse observations. The learned latent prior captures geometry-conditioned variability, enabling rapid and accurate Bayesian inversion without explicit knowledge of the PDEs.
- 3D Car Body Helmholtz Resonance and Source Localization: The posterior over field responses, as well as source localization given partial data, is recovered efficiently via ABC sampling on the geometry-conditioned latent space.
- Complex Terrain Flow: Large-scale multi-GPU training and inference is demonstrated for RANS airflow fields over real-world terrains.
In each case, GABI’s approach yields predictive accuracy comparable to deterministic supervised learning methods in settings where such approaches are feasible, while providing rigorous uncertainty quantification and the flexibility to adapt to arbitrary observation processes (Vadeboncoeur et al., 24 Sep 2025).
6. Comparison to Existing Approaches and Architectural Considerations
GABI’s principal distinction is the decoupling of model learning and inference: prior information is distilled from data irrespective of the future observation process (“train once, use anywhere”). The geometry-aware design contrasts with classical GP latent variable models, which generally assume fixed geometry and may not scale to complex, variable domains. Unlike supervised learning, GABI delivers well-calibrated Bayesian UQ independent of the training task—critical for engineering problems with inherently unpredictable observation structures.
From an architectural perspective, the framework is agnostic: any geometric autoencoder architecture capable of processing (and reconstructing) fields on variable discretizations or graphs is compatible. The method’s reliance on empirical latent space alignment (via divergence matching) and reconstruction error minimization ensures the pushforward prior remains informative and geometry-adaptive (Vadeboncoeur et al., 24 Sep 2025).
7. Significance and Broader Impact
GABI provides a flexible, geometry-adaptive solution to the pervasive challenges of Bayesian inversion in physically complex and data-scarce engineering domains. The “learn first, observe later” separation, geometry-aware generative priors, and sampling-based posterior inference enable principled uncertainty quantification on arbitrary geometries, independent of governing PDEs. The capacity for rapid, robust field reconstruction and UQ underpins potential application in control, monitoring, and diagnostics of physical systems, especially where traditional supervised or analytical inversion methods are intractable or unreliable due to geometric variability and scarce data (Vadeboncoeur et al., 24 Sep 2025).