Shapley Taylor Interaction Index
- STII is a rigorous framework that extends classical Shapley values to attribute higher-order feature interactions.
- It decomposes model outputs into main effects and interactions, uniquely satisfying five axioms for interpretability.
- It offers practical insights for domains like tabular data, NLP, and vision by balancing accuracy with computational scalability.
The Shapley Taylor Interaction Index (STII) is a rigorous, axiomatic framework for attributing the output of a black-box model to feature interactions of arbitrary order. Extending the classical Shapley value, which explains additive feature effects, STII isolates contributions from combinations of features—quantifying higher-order dependencies that are inaccessible to standard single-feature importance metrics. The index is uniquely characterized by an extension of the Shapley axioms and closely parallels the truncated Taylor expansion of a function's multilinear extension, rendering it foundational for interpretable machine learning in domains such as tabular modeling, natural language processing, and vision.
1. Formal Definition and Mathematical Foundation
Let denote the set of features and a value function giving the model output (e.g., probability, logit, or expected prediction) for any subset of “active” features. For an interaction subset , the order- Shapley Taylor Interaction Index is expressed by:
where the discrete derivative is
The weight kernel is
where , . For pairwise interactions, the second-order index isolates non-additive effects between features and . The construction mirrors the Taylor expansion of the multilinear extension of ; main effects are zeroth/first derivatives, and interactions correspond to higher-order discrete differences (Dhamdhere et al., 2019, Hasegawa et al., 5 Dec 2025, Nohara et al., 2022, Singhvi et al., 19 Mar 2024, Bordt et al., 2022).
2. Fundamental Axioms and Uniqueness
STII is determined by five axioms:
- Linearity: for any functions and scalars .
- Dummy: If any feature in is non-interacting (i.e., depends only on ), then all higher-order vanish for .
- Symmetry: Index values are invariant under permutations of feature labels; re-labels accordingly.
- Efficiency: The sum over all subsets of size equals the total gain: .
- Interaction Distribution: Pure higher-order interactions do not “leak” to lower orders; for a pure -way cross function, only is nonzero for .
These axioms, proven in (Dhamdhere et al., 2019, Hasegawa et al., 5 Dec 2025, Bordt et al., 2022), guarantee the uniqueness and interpretive fidelity of STII compared to alternative indices from cooperative game theory.
3. Algorithmic Computation and Scalability
Computation of STII is combinatorial in the number of features. For order- interactions, exact evaluation is per —scaling as for all possible interactions. For tree ensembles, polynomial-arithmetic frameworks such as TreeSHAP-IQ achieve exact indices using recursive traversal with summary, interaction, and quotient polynomials (Muschalik et al., 22 Jan 2024). Permutation-based Monte Carlo approximations, stratified or antithetic sampling, and dynamic programming reduce variance and cost for large-scale problems (Nohara et al., 2022, Singhvi et al., 19 Mar 2024).
Tensor network reformulations (Interaction Tensor SHAP, IT-SHAP) allow for polynomial time and polylog depth computation under Tensor Train decomposition, compressing the exponential complexity of discrete difference enumeration to NC parallel time, conditioned on TT-representability of model and distribution tensors (Hasegawa et al., 5 Dec 2025).
| Method | Order Supported | Complexity |
|---|---|---|
| Brute Force | Any | |
| TreeSHAP-IQ | Any (trees) | |
| Monte Carlo | Any | |
| IT-SHAP | Any |
4. Relationship to Classical Shapley Values and GAMs
The classical Shapley value assigns each feature its marginal contribution, conflating main and interaction effects. STII decomposes the Shapley value into pure main effects (self-interactions) and interactions of arbitrary order. Summing all indices up to order recovers the total model output, satisfying the efficiency axiom (Dhamdhere et al., 2019, Bordt et al., 2022, Nohara et al., 2022).
-order STII recovers Generalized Additive Models (GAMs) with interaction terms up to order : for any , the STII gives exactly (Bordt et al., 2022). This aligns the post-hoc attribution with intrinsic interpretability.
5. Empirical Applications and Case Studies
STII has demonstrated empirical utility across domains:
- Tabular Modeling: In clinical models of colon cancer, STII isolates combinations (e.g., age and bilirubin) whose joint effects exceed main effects—enabling sharper dependence plots and clinicians’ risk stratification (Nohara et al., 2022).
- LLMs: In NLP, pairwise STII quantifies nonlinear syntactic and semantic dependencies (e.g., idiom structure, syntactic proximity). Elevated STII for idiomatic token pairs evidences non-compositionality in model representations (Singhvi et al., 19 Mar 2024).
- Tree Ensembles: Datasets such as German Credit, Bank, or California housing show key features’ importance is supplanted by interactions, e.g., latitude–longitude in real estate (Muschalik et al., 22 Jan 2024).
- Vision and Speech: STII heatmaps delineate object boundaries in images and phonetic coarticulation in speech frames—reflecting structured interaction patterns in model outputs (Singhvi et al., 19 Mar 2024).
6. Practical Considerations and Limitations
Truncation to second- or third-order interactions is typically sufficient in practice for sparse or weakly correlated domains. Computational infeasibility arises for in high dimensions unless leveraging structure (e.g., trees, TT format, dynamic pruning). Sampling variance can yield spurious small interactions. Interventional vs. observational definitions of may alter attribution meaning and computational burden (Muschalik et al., 22 Jan 2024).
Projection onto the most influential interactions, ranking by standard deviation or domain priors, is recommended. Centering interactions (interpreting as the extra contribution beyond main effects) clarifies dependence plots (Nohara et al., 2022, Singhvi et al., 19 Mar 2024).
7. Extensions, Generalizations, and Ongoing Research
The framework admits generalization to structured inputs, as in the Myerson-Taylor index, which internalizes graph connectivity in node-wise and subgraph interactions—uniquely axiomatized for graph neural networks (Bui et al., 23 May 2024). The TT contraction formalism in IT-SHAP unifies main and higher-order effects for scalable interaction analysis in high-dimensional models (Hasegawa et al., 5 Dec 2025).
Recent work connects STII to the broader family of -Shapley values, which characterize all possible function decompositions up to order , bringing post-hoc explanations for complex models into harmony with inherently interpretable model classes (Bordt et al., 2022).
Further methodology refinements center on algorithmic efficiency for large ensembles, tensor network contraction, and background distribution estimation—affecting every facet of practical model interpretation for state-of-the-art black-box architectures.