Deep Gaussian Process Surrogates
- Deep Gaussian Process surrogates are hierarchical models that compose multiple Gaussian process layers to capture nonlinear, nonstationary relationships in data.
- They utilize scalable inference techniques such as inducing points and stochastic expectation propagation to reduce computational cost while maintaining probabilistic accuracy.
- Probabilistic backpropagation enables recursive Gaussian moment matching, providing robust uncertainty quantification crucial for surrogate modeling in complex simulations.
Deep Gaussian Process (DGP) surrogates are a class of surrogate modeling techniques that generalize standard Gaussian processes through hierarchical (multi-layer) functional composition. Each layer in a DGP is a Gaussian process (GP), allowing input–output mappings to be recursively defined as compositions of GPs. This enables the surrogate to capture nonstationary, heteroscedastic, and highly nonlinear dependencies—properties frequently exhibited by complex computer experiments and scientific simulations. DGP surrogates are fully probabilistic, non-parametric, and offer superior capacity for modeling complex relationships and quantifying calibrated predictive uncertainty relative to shallow GPs or deterministic deep neural networks.
1. Hierarchical DGP Framework and Mathematical Structure
A DGP models the data-generating process as a composition of GP layers. For input , the mapping proceeds as:
- Layer GP prior:
- Hidden variable propagation:
with and final predicted output:
Unlike single-layer GPs, this stacked construction allows for automatic learning of input warping, dimensionality expansion or compression, and richer forms of kernel design, all in a data-driven, Bayesian manner. DGPs thereby model complex input–output relationships and uncertainty in high-dimensional, nonstationary settings (Bui et al., 2015).
2. Scalable Inference via Inducing Points and Stochastic Expectation Propagation
Naively, the cost of training a DGP is prohibitive, scaling as for layers and samples. To address this, sparse approximations using a set of inducing points are employed at each GP layer, for example by the Fully Independent Training Conditional (FITC) method:
- Inducing prior:
- Layer transition:
This reduces per-layer training cost to , with inducing points, and enables scalable model training.
Posterior inference in DGPs is analytically intractable, particularly with multiple layers and approximate conditionals. Stochastic Expectation Propagation (SEP) is adopted for scalable approximate Bayesian inference:
where is a Gaussian “average” factor representing a typical data effect, decoupling the memory cost of inference from dataset size. SEP proceeds by sequentially forming cavity distributions, updating with single-likelihood “tilted” moments, and refining . This approach preserves analytic tractability of many updates and enables stochastic optimization with memory cost , independent of (Bui et al., 2015).
3. Probabilistic Backpropagation for Gaussian Moment Matching
A central step in learning and prediction is propagating distributions through the nonlinear hierarchy of GPs. This is most tractable for Gaussians, but deep compositions generally result in non-Gaussian outputs. Probabilistic backpropagation addresses this by, at each layer, matching moments (mean and variance) of the output to a Gaussian, enabling recursive, approximate inference:
- Cavity update:
- Posterior update (for one datapoint, after moment matching):
where is the data likelihood “normalization” under the current approximation, recursively computed via Gaussian marginalization at each DGP layer (Bui et al., 2015).
In two-layer DGPs, integrals such as
are approximated by successive moment matching, using analytic forms for kernels like the exponentiated quadratic under Gaussian input. This makes DGP moment propagation computationally efficient.
4. Advantages in Representational Power and Uncertainty Quantification
DGP surrogates present several key advantages over standard GPs:
- Automated Adaptive Warping: DGPs adaptively re-map the input space, discovering nonlinear transformations appropriate for the data, thereby constructing sophisticated data-driven kernels.
- Expressiveness: The multi-layer structure captures nonstationary and higher-order interactions unreachable by single-layer models.
- Well-Calibrated Uncertainty: Hierarchical composition provides refined uncertainty estimation. By propagating predictive covariance through layers, uncertainty is sensitive to latent representations and modelled function complexity.
- Scalability: Inducing point methods and SEP-based inference enable DGP surrogates to scale to substantial real-world datasets.
- Empirical Performance: On benchmark datasets such as Boston Housing, DGPs trained with SEP and probabilistic backpropagation outperform traditional GP regression in RMSE and mean log loss, and are never worse than state-of-the-art alternatives (Bui et al., 2015).
5. Practical Considerations and Limitations
- Computation: DGPs require more computation than shallow GPs, but SEP reduces memory overhead from (full EP) to . Using minibatch stochastic optimization (e.g., Adam) is essential for large data.
- Inducing Point Placement: Performance benefits from well-placed inducing inputs. Approximations can introduce “approximation repair” effects in DGPs, as higher layers are able to compensate for induced errors at lower layers.
- Approximation Error: Moment matching introduces bias in non-Gaussian regimes, but effectiveness is sustained in practical datasets using analytic moment propagation for common kernels.
- Depth: Empirically, two or three layers often suffice; deeper structures can be more difficult to train reliably.
6. Applications, Impact, and Related Developments
DGP surrogates are widely applicable where flexibility, nonstationarity, and robust uncertainty quantification are crucial. Examples include:
- Scientific and Engineering Simulations: Surrogate modeling of computationally intensive simulators with complex, non-stationary response surfaces.
- Automated Bayesian Kernel Design: Data-driven learning of warping and expansion/compression avoids manual feature engineering.
- Active Learning and Bayesian Optimization: DGP uncertainty quantification informs acquisition strategies in sequential experimental design settings.
Related advancements leverage DGP frameworks in diverse inferential settings—for instance, Bayesian inverse problems, active learning with complex simulators, and multi-task modeling—due to their capacity for both flexible function approximation and rigorous uncertainty modeling (Bui et al., 2015).
In summary, Deep Gaussian Process surrogates, as realized through sparse inducing point approximations, stochastic expectation propagation, and probabilistic backpropagation, offer a tractable, scalable Bayesian surrogate modeling solution that flexibly adapts to complex, non-stationary relationships and provides high-quality uncertainty estimates, outperforming conventional Gaussian Process regression across benchmark tasks.