Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 128 tok/s
Gemini 2.5 Pro 44 tok/s Pro
GPT-5 Medium 28 tok/s Pro
GPT-5 High 23 tok/s Pro
GPT-4o 75 tok/s Pro
Kimi K2 189 tok/s Pro
GPT OSS 120B 432 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Functional ANOVA Decomposition in RKHS

Updated 2 October 2025
  • Functional ANOVA decomposition is a method that represents a multivariate function as an orthogonal sum of contributions from individual variables and their interactions.
  • It leverages reproducing kernel Hilbert spaces and centered ANOVA kernels to obtain explicit, non-recursive decompositions and analytical sensitivity index computations.
  • This approach underpins surrogate modeling, Gaussian process regression, and global sensitivity analysis, offering clarity and computational efficiency in high-dimensional settings.

Functional ANOVA decomposition is a mathematical formulation that expresses a multivariate function as an orthogonal sum of components associated with individual variables and their interactions. This construction has deep connections to variance-based sensitivity analysis, kernel methods, statistical learning, and interpretability in machine learning. The RKHS-centric framework described in (Durrande et al., 2011) provides foundational advances in the theory and practical computation of functional ANOVA decompositions, especially as they relate to kernel interpolation, Gaussian process regression, and global sensitivity analysis.

1. Functional ANOVA Decomposition: Foundational Concepts

Functional ANOVA (Analysis of Variance) decomposition represents a real-valued function f:DRdRf: D \subseteq \mathbb{R}^d \to \mathbb{R} as a sum of terms reflecting the contribution of individual variables and subsets of variables to the output. More formally, under suitable integrability and measure-theoretic assumptions, the decomposition can be written as: f(x)=f0+i=1dfi(xi)+1i<jdfij(xi,xj)++f1d(x1,...,xd)f(x) = f_0 + \sum_{i=1}^d f_i(x_i) + \sum_{1 \leq i < j \leq d} f_{ij}(x_i, x_j) + \cdots + f_{1\cdots d}(x_1,...,x_d) where f0f_0 is a constant, fif_i are main effects, fijf_{ij} are second-order interactions, etc. Each component is orthogonal in L2(μ)L^2(\mu) under the product measure μ\mu and satisfies suitable centering constraints (e.g., main effects have zero mean with respect to the marginal of xix_i).

The RKHS approach in (Durrande et al., 2011) establishes a rigorous pathway to constructing function spaces and kernels wherein this decomposition is not only possible, but is mirrored algebraically by the kernel itself. This connection is central to ensuring both orthogonality and easy computation of sensitivity indices.

2. RKHS Decomposition: Zero-Mean Functions and Kernel Splitting

Let H\mathcal{H} be a one-dimensional RKHS with reproducing kernel kk over domain DD and measure μ\mu. The decomposition leverages the integral operator I:hDh(s)dμ(s)I : h \mapsto \int_D h(s) d\mu(s) and the Riesz representation theorem to write: H=H0H1\mathcal{H} = \mathcal{H}_0 \oplus^\perp \mathcal{H}_1 where:

  • H0={hH:Dh(s)dμ(s)=0}\mathcal{H}_0 = \left\{ h \in \mathcal{H} : \int_D h(s)d\mu(s) = 0 \right\} is the subspace of zero-mean functions (centered w.r.t. μ\mu)
  • H1=span{R}\mathcal{H}_1 = \text{span}\{R\}, with RR the representer of the integral

Correspondingly, the kernel itself splits as: k(x,y)=k0(x,y)+k1(x,y)k(x, y) = k_0(x, y) + k_1(x, y) with: k0(x,y)=k(x,y)Dk(x,s)dμ(s)Dk(y,s)dμ(s)D×Dk(s,t)dμ(s)dμ(t)k_0(x, y) = k(x, y) - \frac{\int_D k(x, s)d\mu(s) \int_D k(y, s)d\mu(s)}{\iint_{D \times D} k(s, t)d\mu(s)d\mu(t)} The function xk0(x,y)x \mapsto k_0(x, y) is centered for each yy, i.e., Dk0(x,y)dμ(x)=0\int_D k_0(x, y)d\mu(x) = 0.

In high dimensions, these zero-mean subspaces are combined using tensor products and correspond to subcomponents (main effects, interactions) in the ANOVA decomposition. This approach is the foundation for the construction of “centered” ANOVA (K*_ANOVA) kernels.

3. Construction of (Centered) ANOVA Kernels

Given univariate kernels kik^i (with their decomposition into ki=k0i+k1ik^i = k^i_0 + k^i_1), the multivariate (centered) ANOVA kernel is defined as: KANOVA(x,y)=i=1d[1+k0i(xi,yi)]K^*_{\text{ANOVA}}(x, y) = \prod_{i=1}^d [1 + k^i_0(x_i, y_i)] which expands as: KANOVA(x,y)=1+I{1,...,d}iIk0i(xi,yi)K^*_{\text{ANOVA}}(x, y) = 1 + \sum_{I \subseteq \{1, ..., d\}} \prod_{i \in I} k^i_0(x_i, y_i) This kernel's RKHS is composed of functions whose ANOVA terms (main and interactions) are zero-mean and mutually orthogonal in L2(μ)L^2(\mu). This property is critical for the interpretability and computability of the corresponding decomposition and underpins the construction of surrogate models in computer experiments.

4. Analytical Formulas for RKHS Best Predictors and Decomposition

When building the best predictor (either via interpolation or regularization) in the RKHS with K*_ANOVA, the function admits the decomposition: m(x)=m0+i=1dmi(xi)+i<jmij(xi,xj)++m1...d(x)m(x) = m_0 + \sum_{i=1}^d m_i(x_i) + \sum_{i < j} m_{ij}(x_i, x_j) + \cdots + m_{1...d}(x) with

mI(x)=(iIk0i(xi))TK1Fm_I(x) = \left( \bigodot_{i \in I} k^i_0(x_i) \right)^T K^{-1} F

where \odot denotes elementwise product across observed design points, KK is the Gram matrix, and FF is the vector of observations. This formula is notable for being non-recursive and mirror-symmetric to the functional ANOVA decomposition of the true function.

5. Sensitivity Analysis and Computation of Sobol’ Indices

The decomposition of mm yields mutually orthogonal ANOVA submodels, allowing a variance-based breakdown: Var(m(X))=I{1,...,d}Var(mI(XI))\operatorname{Var}(m(X)) = \sum_{I \subseteq \{1,...,d\}} \operatorname{Var}( m_I( X_I ) ) with corresponding Sobol indices: SI=Var(mI(XI))Var(m(X))S_I = \frac{ \operatorname{Var}(m_I(X_I)) }{ \operatorname{Var}(m(X)) } A distinguishing contribution of (Durrande et al., 2011) is that these sensitivity indices can be computed analytically using explicit formulas involving the kernel integrals: SI=FTK1(iIΓi)K1FFTK1{i=1d(1n×n+Γi)1n×n}K1FS_I = \frac{ F^T K^{-1} \left( \bigodot_{i \in I} \Gamma_i \right) K^{-1} F }{ F^T K^{-1} \left\{ \bigodot_{i=1}^d ( 1_{n \times n} + \Gamma_i ) - 1_{n \times n} \right\} K^{-1} F } with

Γi=Dik0i(x)k0i(x)Tdμ(x)\Gamma_i = \int_{D_i} k^i_0( x ) k^i_0( x )^T d\mu( x )

This closed-form computation—avoiding recursive integration—enables efficient and numerically stable estimation of variable importance in high-dimensional models.

6. Practical Applications and Case Studies

The RKHS-based functional ANOVA decomposition is especially suited for:

  • Surrogate modeling of expensive simulators in engineering and physical sciences, where only a limited number of runs are feasible, and global interpretation in terms of variable effects is critical.
  • Gaussian process regression (kriging) with ANOVA or K*_ANOVA kernels, ensuring each effect is interpretable and orthogonal, with main and interaction terms clearly isolated.
  • Analytical and non-recursive computation of global sensitivity indices, critical for model simplification, variable screening, and understanding uncertainty contributions.

The methodology is validated on canonical test functions (e.g., Sobol g-function) and other benchmark problems, where the numerical results confirm the accurate isolation and quantification of main effects and higher-order interactions.

7. Mathematical and Algorithmic Framework

Key mathematical tools:

  • Integral operator theory: enabling the identification of zero-mean subspaces and the construction of representers within RKHS.
  • Tensor product structure of kernels: facilitating multivariate decomposition as a sum over all subsets, aligned with the functional ANOVA form.
  • Closed-form interpolant and regularization solutions: allowing for both exact and regularized fits in the RKHS, with the regularization parameter λ\lambda controlling noise levels.

Key algorithmic steps involve:

  • Construction of centered univariate kernels k0ik^i_0 using kernel integral formulas.
  • Assembly of multivariate K*_ANOVA kernel.
  • Computation of Gram matrices and model predictions using powers of kernel integrals.
  • Direct computation of sensitivity indices from fitted model coefficients, circumventing recursion.

8. Impact and Implications

The RKHS-centered approach to functional ANOVA decomposition in (Durrande et al., 2011) achieves:

  • Orthogonal, interpretable, and non-recursive decomposition of fitted models.
  • Efficient, closed-form variance and sensitivity index computation.
  • Applicability to high-stakes domains (scientific, engineering) where model evaluation is costly and variable prioritization is essential.
  • Seamless integration into existing kernel-based nonparametric regression frameworks.

This theoretical and practical framework has become foundational in surrogate modeling, global sensitivity analysis, and the interpretable design of kernel-based statistical learning models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Functional ANOVA Decomposition.