Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 128 tok/s
Gemini 2.5 Pro 44 tok/s Pro
GPT-5 Medium 28 tok/s Pro
GPT-5 High 23 tok/s Pro
GPT-4o 75 tok/s Pro
Kimi K2 189 tok/s Pro
GPT OSS 120B 432 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Generalized Hoeffding Decomposition

Updated 9 October 2025
  • Generalized Hoeffding Decomposition is an extension of the classical technique that decomposes multivariate functions with arbitrarily dependent inputs into hierarchically orthogonal components.
  • It employs projection operators and conditional expectations to separate variance into structural and correlative contributions, facilitating accurate global sensitivity analysis.
  • This framework enhances model interpretability and feature selection in dependent input scenarios, leveraging iterative methods and nonparametric estimation for practical analysis.

The Generalized Hoeffding Decomposition (GHD) extends the classical orthogonal expansion of square-integrable functions of multiple random inputs to cases in which the input variables exhibit arbitrary dependencies. Traditional Hoeffding decompositions, foundational in probabilistic analysis and sensitivity quantification, rely on mutual independence of inputs to guarantee mutual orthogonality of summands corresponding to different variable coalitions. GHD provides a rigorous framework for decomposing functions of dependent inputs under mild regularity and boundedness conditions on the joint distribution, resulting in hierarchically orthogonal expansions and a variance attribution scheme that incorporates both structural (variance) and correlative (covariance) contributions. This formalism underpins state-of-the-art approaches to global sensitivity analysis and model interpretability when dependencies are present.

1. Classical Versus Generalized Hoeffding Decomposition

In the classical setting, for a function η:RpR\eta: \mathbb{R}^p \rightarrow \mathbb{R} defined on independent random inputs X=(X1,...,Xp)X = (X_1, ..., X_p), the Hoeffding decomposition takes the form: η(x)=η0+i=1pηi(xi)+1i<jpηij(xi,xj)++η1,,p(x1,,xp)\eta(x) = \eta_0 + \sum_{i=1}^p \eta_i(x_i) + \sum_{1 \leq i < j \leq p} \eta_{ij}(x_i, x_j) + \cdots + \eta_{1,\dots,p}(x_1,\dots,x_p) where each component function is mutually orthogonal with respect to the product measure induced by independence; hence, the variance of the output splits additively: Var(Y)=u{1,,p}Var(ηu(Xu))\mathrm{Var}(Y) = \sum_{u \subseteq \{1,\dots,p\}} \mathrm{Var}(\eta_u(X_u)) Sobol indices Su=Var(ηu(Xu))/Var(Y)S_u = \mathrm{Var}(\eta_u(X_u))/\mathrm{Var}(Y) quantify the sensitivity of each coalition.

When inputs are dependent, the classical mutually orthogonal subspaces do not exist. The generalized framework, under the assumption that PXP_X is absolutely continuous with respect to a product measure ν\nu and satisfies a lower-bounded density condition

pX(x)MpXu(xu)pXuc(xuc),0<M1,p_X(x) \geq M\cdot p_{X_u}(x_u)p_{X_{u^c}}(x_{u^c}), \qquad 0 < M \leq 1,

constructs decomposition into subspaces Hu0H_u^0 of L2L^2 functions of XuX_u orthogonal to all lower-order effects, leading to the unique expansion: η(X)=u{1,,p}ηu(Xu),ηuHu0.\eta(X) = \sum_{u \subseteq \{1,\dots,p\}} \eta_u(X_u), \quad \eta_u \in H_u^0. Orthogonality is now hierarchical: ηu,ηv=0\langle \eta_u, \eta_v \rangle = 0 if vuv \subsetneq u, but not, in general, for unordered pairs.

2. Decomposition and Projection Operators

For the generalized setting, the decomposition is constructed via projectors. For two variables, the system reads: Y=η(X1,X2)=η0+η1(X1)+η2(X2)+η12(X1,X2)Y = \eta(X_1, X_2) = \eta_0 + \eta_1(X_1) + \eta_2(X_2) + \eta_{12}(X_1, X_2) with projections explicitly given by conditional expectation operators:

  • PH(η)=E[η(X)]P_{H_\emptyset}(\eta) = \mathbb{E}[\eta(X)]
  • PH10(η)=E[η(X)X1]E[η(X)]P_{H_1^0}(\eta) = \mathbb{E}[\eta(X) | X_1] - \mathbb{E}[\eta(X)]
  • PH20(η)=E[η(X)X2]E[η(X)]P_{H_2^0}(\eta) = \mathbb{E}[\eta(X) | X_2] - \mathbb{E}[\eta(X)]
  • PH120(η)=η(X1,X2)E[ηX1]E[ηX2]+E[η]P_{H_{12}^0}(\eta) = \eta(X_1, X_2) - \mathbb{E}[\eta | X_1] - \mathbb{E}[\eta | X_2] + \mathbb{E}[\eta]

Solving for the component functions is operationally mapped to solving a linear system involving these special projectors, which may be numerically addressed via iterative methods such as Gauss–Seidel, particularly when symmetry or sparsity is present.

3. Hierarchical Orthogonality and Sensitivity Indices

Because the decomposed components are not generally mutually orthogonal, additional cross-terms enter the variance decomposition: Var(Y)=u[Var(ηu(Xu))+vu,uvCov(ηu(Xu),ηv(Xv))]\mathrm{Var}(Y) = \sum_{u \not= \emptyset} \left[ \mathrm{Var}(\eta_u(X_u)) + \sum_{v \not= u, u\cap v \neq \emptyset} \mathrm{Cov}(\eta_u(X_u), \eta_v(X_v)) \right] The generalized sensitivity index for set uu is given as

Su=Var(ηu(Xu))+v:uvCov(ηu(Xu),ηv(Xv))Var(Y)S_u = \frac{ \mathrm{Var}(\eta_u(X_u)) + \sum_{v: u\cap v \neq \emptyset} \mathrm{Cov}(\eta_u(X_u), \eta_v(X_v)) }{\mathrm{Var}(Y)}

These indices sum to unity even when input variables are dependent: u{1,...,p},uSu=1\sum_{u \subseteq \{1,...,p\}, u \neq \emptyset} S_u = 1 The structure delineates the "structural" (variance) from "correlative" (covariance) contributions of each input coalition to output variability—a necessity for meaningful sensitivity analysis in dependent scenarios.

4. Assumptions, Copula Equivalence, and Estimation Algorithms

The GHD requires:

  • Absolute continuity of joint law PXP_X with respect to ν\nu, typically a product measure.
  • Lower-bounded density condition as described above. For two inputs, a related condition on the copula density is equivalent to the required product-form bound.

Estimation in practice entails nonparametric regression for conditional expectations—a kernel-based local polynomial regression with leave-one-out cross-validation is suggested. Matrix inversion in local polynomial estimation is efficiently handled via the Sherman–Morrison formula. For functional equations arising from the projection system, an iterative numerical Gauss–Seidel algorithm is effective.

In the case where inputs are grouped into independent pairs with within-pair dependence (IPDV models), the decomposition applies first at the group level (classical orthogonal decomposition), then at the pair level (independent pairwise generalized decomposition), allowing refined attribution via first and second-order indices.

5. Practical Sensitivity Analysis and Interpretability

The GHD, when combined with new indices, provides interpretable sensitivity analysis of model outputs in non-independent regimes, now standard in realistic scenarios. It quantifies both direct and indirect effects of input variables, capturing how dependency inflates or modulates the contribution of input coalitions.

For model calibration, reliability, or reverse engineering, the separation into variance and covariance contributions immediately identifies structurally influential inputs and disentangles these from purely correlative effects, enhancing the decision-support robustness.

The indices allow resource allocation, feature selection, or uncertainty quantification strategies to correctly account for interdependence, preventing misattribution of importance that can occur when naively assuming independence.

6. Extensions, Limitations, and Future Directions

The framework is robust for distributions admitting densities with respect to product measures with boundedness as required. For higher dimensions, the approach generalizes naturally via recursive projectors onto hierarchically orthogonal subspaces.

For estimation, nonparametric procedures scale with sample size and variable count; improvements in computational linear algebra and sample-efficient regressors further ameliorate resource requirements.

Open problems include relaxing density assumptions, addressing severely degenerate or highly sparse joint supports, and extending the method to discrete or categorical input spaces beyond the binary case, e.g., via combinatorial or algebraic modifications to the projection system.

Finally, the approach bridges classical ANOVA-type decompositions and modern explainable machine learning needs for "black-box" models, incorporating dependence into attribution and certification frameworks. This positions GHD as a core methodology for interpretable statistical analysis in complex, correlated input domains (Chastaing et al., 2011).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Generalized Hoeffding Decomposition.