Papers
Topics
Authors
Recent
Search
2000 character limit reached

FANOVA-GP Prior: Sensitivity Analysis Framework

Updated 2 May 2026
  • The paper introduces the FANOVA-GP prior, an orthogonal GP framework that decomposes functions into additive effects for sensitivity analysis.
  • The methodology constructs effect-specific kernels through conditional orthogonalization, enabling scalable computation and clear interpretability.
  • It provides analytic Sobol’ and Shapley indices for feature attributions, demonstrating effectiveness in high-dimensional and functional output scenarios.

The FANOVA-GP prior (Functional ANOVA Gaussian Process prior) provides a nonparametric Bayesian framework for variance-based sensitivity analysis and effect decomposition in computer experiments, particularly targeting problems involving functional or high-dimensional outputs. By enforcing explicit orthogonality among all additive and interaction effects, it generalizes the classical ANOVA decomposition to arbitrary distributions, nonlinear dependencies, and both scalar and functional responses. Key constructions include conditional-orthogonality via kernel conditioning, closed-form effect-specific kernels, and efficient algorithms for both effect separation and attributions such as Sobol' and Shapley indices.

1. Mathematical Foundations of the FANOVA-GP Prior

The FANOVA-GP prior models the latent function f(x)f(x), with input x=(x1,,xd)x = (x_1, \dots, x_d), as an explicit additive decomposition: f(x)=SDfS(xS),f(x) = \sum_{S \subseteq D} f_S(x_S), where D={1,,d}D = \{1, \dots, d\}, SS indexes any subset of features, and xSx_S denotes the subvector corresponding to SS. Each component fSf_S is given a zero-mean Gaussian process prior independent of other components: fSGP(0,KS(xS,xS)),f_S \sim \mathcal{GP}\bigl(0, K_S(x_S, x_S')\bigr), ensuring orthogonality of effects under the feature-space measure: Cov(fS(xS),fT(xT))=0(ST).\operatorname{Cov}(f_S(x_S), f_T(x_T')) = 0 \qquad (S \ne T). The aggregate prior thus corresponds to a GP with kernel

x=(x1,,xd)x = (x_1, \dots, x_d)0

In the case of functional outputs, a joint variable x=(x1,,xd)x = (x_1, \dots, x_d)1 is considered, e.g.,

x=(x1,,xd)x = (x_1, \dots, x_d)2

and the decomposition becomes

x=(x1,,xd)x = (x_1, \dots, x_d)3

where each x=(x1,,xd)x = (x_1, \dots, x_d)4 is a zero-mean GP with covariance x=(x1,,xd)x = (x_1, \dots, x_d)5 and conditional independence among components (Tan et al., 15 Jun 2025, Mohammadi et al., 20 Aug 2025).

2. Kernel Construction and Conditional-Orthogonality

Orthogonality of components is enforced at the kernel level. For each feature x=(x1,,xd)x = (x_1, \dots, x_d)6, a base positive-definite kernel x=(x1,,xd)x = (x_1, \dots, x_d)7 (e.g., squared-exponential) is orthogonalized to yield a zero-mean kernel: x=(x1,,xd)x = (x_1, \dots, x_d)8 Each interaction kernel for subset x=(x1,,xd)x = (x_1, \dots, x_d)9 is then constructed as

f(x)=SDfS(xS),f(x) = \sum_{S \subseteq D} f_S(x_S),0

The full kernel is the sum over all f(x)=SDfS(xS),f(x) = \sum_{S \subseteq D} f_S(x_S),1.

For functional-output decompositions, the output domain kernel f(x)=SDfS(xS),f(x) = \sum_{S \subseteq D} f_S(x_S),2 is introduced, and the subsetwise kernel is

f(x)=SDfS(xS),f(x) = \sum_{S \subseteq D} f_S(x_S),3

Orthogonality is guaranteed by conditioning the GP prior for f(x)=SDfS(xS),f(x) = \sum_{S \subseteq D} f_S(x_S),4 on the constraint that its mean under each f(x)=SDfS(xS),f(x) = \sum_{S \subseteq D} f_S(x_S),5 marginal is zero for every f(x)=SDfS(xS),f(x) = \sum_{S \subseteq D} f_S(x_S),6 and for all f(x)=SDfS(xS),f(x) = \sum_{S \subseteq D} f_S(x_S),7: f(x)=SDfS(xS),f(x) = \sum_{S \subseteq D} f_S(x_S),8 This is achieved analytically through Gaussian conditioning, where the same kernel formula arises from the removal of the mean component. Consequently, all effect components are mutually orthogonal in f(x)=SDfS(xS),f(x) = \sum_{S \subseteq D} f_S(x_S),9 for the empirical feature distribution, and each effect has zero expectation in each coordinate (Tan et al., 15 Jun 2025, Mohammadi et al., 20 Aug 2025).

3. Analytical Indices for Sensitivity Analysis

Once the posterior GP is fitted, one obtains a posterior mean function decomposed by effect: D={1,,d}D = \{1, \dots, d\}0 with each D={1,,d}D = \{1, \dots, d\}1 computable via closed-form kernel evaluations and GP weights.

Variance-based sensitivity analysis is realized by computing, for each effect D={1,,d}D = \{1, \dots, d\}2 and output location D={1,,d}D = \{1, \dots, d\}3, the local variance: D={1,,d}D = \{1, \dots, d\}4 and the total local variance D={1,,d}D = \{1, \dots, d\}5, with the local Sobol' index: D={1,,d}D = \{1, \dots, d\}6 Closed-form expressions for D={1,,d}D = \{1, \dots, d\}7 are available: D={1,,d}D = \{1, \dots, d\}8 where D={1,,d}D = \{1, \dots, d\}9 and SS0 denotes empirical expectations over the SS1 marginal (Tan et al., 15 Jun 2025).

Averaging these indices in SS2 yields global (expected conditional variance, ECV) variances and ECV indices: SS3 enabling comprehensive attribution of both main effects and interactions.

4. Efficient Computation and Inference Procedures

The computational workflow closely follows that of standard GP regression, with added steps for empirical integration and kernel assembly:

  1. Empirical estimation of marginal feature densities SS4 (or use of observed marginals) for construction of each SS5.
  2. Assembly of the additive, orthogonal kernel matrix SS6 via Hadamard products or, for the additive representation, via Newton's identities for elementary symmetric polynomials, leading to SS7 complexity per kernel evaluation.

The covariance matrix in the functional-output case is

SS8

with SS9, xSx_S0. The marginal likelihood is optimized w.r.t. hyperparameters (scales xSx_S1, noise xSx_S2, length scales) using gradient-based solvers, and xSx_S3 has a closed-form maximizer.

If outputs are observed on a regular grid in xSx_S4 and xSx_S5, the full kernel decomposes as a Kronecker product, yielding xSx_S6 cost for xSx_S7 grids (Tan et al., 15 Jun 2025).

The implementation for non-functional outputs benefits from explicit Möbius inversion and recursion over symmetric polynomials, avoiding enumeration of all xSx_S8 subsets, with all key operations scaling quadratically in xSx_S9 (Mohammadi et al., 20 Aug 2025).

5. Shapley Attributions and Explainability

The FANOVA-GP family admits exact, closed-form computation of Shapley values for both local (instance-wise) and global (variance-based) feature attributions at quadratic time complexity. The stochastic Shapley value is defined via the cooperative game over function components, capturing the expected contribution (as well as uncertainty) of each input to the functional or scalar output. Global Shapley values quantify feature importance for the model's overall sensitivity structure.

These attributions rest on a Möbius representation of the FANOVA decomposition and recursive algorithms leveraging Newton's identities for elementary symmetric polynomials, facilitating scalable and axiomatically sound explainability for structured probabilistic models (Mohammadi et al., 20 Aug 2025).

6. Nonparametric, Data-Driven, and Orthogonal Properties

No fixed basis functions are required, and the orthogonality constraint is imposed analytically in the kernel. This enables:

  • Fully nonparametric modeling, with data-driven orthogonality valid for any observed SS0 marginal distribution.
  • No strong distributional assumptions, as all required kernel integrals are approximated empirically from the data.
  • All effect orders (including high-order interactions) present in the prior, controlled by separate scale hyperparameters; uninformative high-order terms are naturally shrunk by their learned variances.
  • Equivalence of computational costs to ordinary GP regression, with only minor preprocessing for empirical integral estimation.

This approach yields an explicit, orthogonal, and interpretable ANOVA decomposition for complex, nonlinear, and functional-output computer experiments, obviating the need for manual basis selection or uniform input assumptions, and with analytic variance-based indices available without resorting to Monte Carlo estimation (Tan et al., 15 Jun 2025).

7. Practical Relevance and Applications

The practical utility of FANOVA-GP priors, including the FOAGP variant for functional outputs, is demonstrated by their effective orthogonal effect decomposition and variance analysis in both simulated and real engineering applications, such as fuselage shape control. This framework provides both practitioners and theorists with an analytically tractable, scalable, and robust tool for nonparametric sensitivity analysis, interpretable uncertainty quantification, and input attribution in a wide range of complex, black-box modeling scenarios (Tan et al., 15 Jun 2025, Mohammadi et al., 20 Aug 2025).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to FANOVA-GP Prior.