Sparse Isotonic Shapley Regression (SISR)
- SISR is a framework that restores additivity in Shapley values through a learned monotonic transformation of coalition payoffs.
- It alternates between isotonic regression and normalized hard-thresholding to enforce exact ℓ0 sparsity in attributions.
- Empirical evaluations show that SISR delivers stable and interpretable explanations even with heavy-tailed, dependent features.
Sparse Isotonic Shapley Regression (SISR) addresses two foundational obstacles in Shapley-based model explainability: the violation of the canonical additivity assumption in real-world payoff constructions, and the need for sparse, interpretable attributions in high-dimensional feature spaces. The method provides a unified framework that simultaneously estimates a monotonic, data-driven transformation to restore additivity of the worth function and enforces an sparsity constraint on the Shapley vector, thereby yielding more consistent and computationally efficient explanations in settings plagued by non-Gaussianity, heavy tails, feature dependence, or idiosyncratic loss scales (She, 2 Dec 2025).
1. Model Formulation and Objectives
SISR considers a family of coalition values and seeks to jointly: (i) learn a strictly increasing scalar function with , such that becomes additive in the underlying Shapley scores; (ii) enforce sparsity ( constraint) on the Shapley vector , aligning with the hypothesis that only a small subset of features is truly relevant.
The model assumes that for all coalitions ,
Enforcing directly achieves exact sparsity in attributions, bypassing the shrinkage-bias inherent to penalization. The primary SISR objective is:
Reparameterization with and (discrete), and introducing matrix (rows index coalitions), yields the equivalent quadratic problem with order restrictions on .
2. Optimization Algorithm
SISR employs alternating minimization over (sparse Shapley vector) and (discretized monotonic transform):
A. Monotonic Transformation via Isotonic Regression:
With fixed, update by solving a weighted isotonic regression problem:
This is executed exactly using the Pool-Adjacent-Violators (PAV) algorithm with computational complexity , for (often reduced via subsampling). The output vector provides the transformed payoffs at observed coalition values.
B. Sparsity via Normalized Hard-Thresholding:
With fixed, update by minimizing the objective under constraints. The step utilizes a quadratic surrogate and a hard-thresholding operator , where zeros all but the largest . The global minimizer takes the form:
where , .
Algorithm Sketch:
- Initialize as unit-norm -sparse vector, ; precompute .
- Alternate: (a) update by normalized hard-thresholding; (b) update with weighted PAVA.
- Continue until objective stabilizes.
Convergence properties are established: each -update reduces objective, the sequence converges, and gaps vanish if is strictly greater than .
3. Theoretical Guarantees and Analysis
Under the generative model , with -sparsity of and , the following results are obtained:
- Consistency: As noise vanishes or sample size of coalition values increases, the isotonic regression step recovers the true monotonic transformation up to negligible error.
- Support Recovery: Provided the minimum nonzero is substantially larger than the noise and the signal-to-noise ratio is large, the normalized hard-thresholding step identifies the correct support with high probability.
- Feature Dependence and Heavy-Tailed Payoffs: SISR learns a transform that absorbs the nonlinearities, restoring additivity even for payoff structures arising from feature dependence, heavy tails, or domain-specific loss constructions (e.g., under correlated regressors). The canonical linear-Gaussian assumption underlying standard Shapley can otherwise suffer severe violations leading to distorted attributions.
4. Empirical Performance
Comprehensive experiments on synthetic and real data characterize SISR performance:
- Synthetic Recovery: For a battery of toy monotonic transforms (e.g., square root, fifth root, logarithm, exponential, tanh-like, Gaussian CDF-odds), SISR accurately fits (correlation ).
- Sparse Support Recovery: For synthetic with , up to 25, and noise up to , both support agreement and affinity remain near 1. Recovery degrades gracefully with noise; e.g., for , true support is recovered approximately 90% of the time.
- Payoff Scenarios: Simulations with payoff under feature correlation/irrelevance demonstrate that raw is not additive; SISR fits a correction, delivering stable attributions, while standard Shapley responses become highly nonrobust.
- Real Data Applications:
- Prostate dataset: Shapley values exaggerate the importance of “svi”; SISR zeros it out, aligning with AIC/BIC and expert knowledge.
- Boston housing: With robust payoff , baseline Shapley values flip important variable signs; SISR’s log-like transform yields stable rankings.
- Credit and Pima diabetes: Across diverse payoff schemes, SISR provides consistent sparse attributions (selected via RIC), whereas standard SAGE/Shapley attributions are unstable.
5. Computational and Practical Considerations
- Complexity: Weighted PAVA is per iteration. When utilizing sparse storage for (each coalition row nonzeros), -updates cost per iteration. Convergence is typically reached within tens of alternations.
- Sparsity Level Selection: If is unknown, Risk Inflation Criterion (RIC) can be used to choose . Empirical evidence indicates reliable selection.
- Initialization: as the top- entries of (normalized) and (with large for numerical stability) are recommended.
- Infinite Weights for Special Coalitions: To ensure and the “efficiency” constraint, theoretical Shapley weights and diverge; in practice, substituting large multiples (e.g., 10 or 100) of the maximum weight is effective.
- Extensibility: Alternating PAVA plus hard-thresholding can extend to other convex loss functions (e.g., GLMs with canonical links) by substituting the quadratic surrogate with a suitable local approximation.
6. Significance and Implications in Explainable AI
SISR advances the Shapley explanation paradigm by jointly resolving two stated limitations: First, it restores additivity in the presence of nonadditive, domain-specific, or statistically problematic payoff functions via a learned monotonic transformation. Second, it provides exact -sparse attributions in high-dimensional settings, overcoming the inconsistency and computational burden of post hoc thresholding.
Empirical evidence demonstrates that SISR explanations are both more stable and faithful to domain structure—eliminating attributions to irrelevant features and correcting severe rank/sign distortions observed in conventional Shapley or SAGE values under nonlinear or unstable payoffs.
A plausible implication is that SISR is well-positioned as a general-purpose attribution framework for scenarios involving heavy tails, feature dependence, or overarching nonlinearity in payoff structure, where standard linear Shapley values are provably unreliable (She, 2 Dec 2025).