Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 73 tok/s
Gemini 2.5 Pro 51 tok/s Pro
GPT-5 Medium 31 tok/s Pro
GPT-5 High 32 tok/s Pro
GPT-4o 103 tok/s Pro
Kimi K2 218 tok/s Pro
GPT OSS 120B 460 tok/s Pro
Claude Sonnet 4.5 35 tok/s Pro
2000 character limit reached

Disentangled Feature Importance (2507.00260v1)

Published 30 Jun 2025 in stat.ML, cs.LG, math.ST, stat.ME, and stat.TH

Abstract: Feature importance quantification faces a fundamental challenge: when predictors are correlated, standard methods systematically underestimate their contributions. We prove that major existing approaches target identical population functionals under squared-error loss, revealing why they share this correlation-induced bias. To address this limitation, we introduce \emph{Disentangled Feature Importance (DFI)}, a nonparametric generalization of the classical $R2$ decomposition via optimal transport. DFI transforms correlated features into independent latent variables using a transport map, eliminating correlation distortion. Importance is computed in this disentangled space and attributed back through the transport map's sensitivity. DFI provides a principled decomposition of importance scores that sum to the total predictive variability for latent additive models and to interaction-weighted functional ANOVA variances more generally, under arbitrary feature dependencies. We develop a comprehensive semiparametric theory for DFI. For general transport maps, we establish root-$n$ consistency and asymptotic normality of importance estimators in the latent space, which extends to the original feature space for the Bures-Wasserstein map. Notably, our estimators achieve second-order estimation error, which vanishes if both regression function and transport map estimation errors are $o_{\mathbb{P}}(n{-1/4})$. By design, DFI avoids the computational burden of repeated submodel refitting and the challenges of conditional covariate distribution estimation, thereby achieving computational efficiency.

Summary

  • The paper introduces DFI to overcome the limitations of conventional feature attribution methods under correlated predictors.
  • It establishes a robust theoretical framework with root-n consistency and asymptotic normality for reliable importance estimation.
  • Empirical results show that DFI recovers true importance structures efficiently, even in high-dimensional and strongly correlated scenarios.

Disentangled Feature Importance: A Principled Approach to Feature Attribution under Correlation

The paper "Disentangled Feature Importance" (2507.00260) addresses a central challenge in interpretable machine learning: quantifying feature importance in the presence of correlated predictors. The authors rigorously demonstrate that widely used methods—such as LOCO, conditional permutation importance (CPI), and Shapley values—are fundamentally limited by their inability to properly account for feature dependencies. This limitation leads to systematic underestimation or misattribution of importance, particularly in high-correlation regimes, and is shown to be intrinsic to the population functionals these methods target.

Theoretical Foundations and Critique of Existing Methods

The paper establishes, via elementary probability and variance decomposition, that LOCO and CPI both estimate the same conditional variance functional under squared-error loss:

ψXj=E[Var(μ(X)Xj)]\psi_{X_j} = \mathbb{E}\left[\mathrm{Var}(\mu(X) \mid X_{-j})\right]

where μ(X)\mu(X) is the regression function. This equivalence, while elegant, exposes a critical flaw: when features are perfectly or near-perfectly correlated, the conditional variance collapses, and genuinely influential features can receive zero importance. The authors provide explicit examples where, despite strong dependence of YY on XjX_j, both LOCO and CPI assign zero importance to XjX_j if it is deterministically related to other features. The Shapley value, as a weighted average of LOCO scores across all submodels, inherits this pathology in high dimensions.

Disentangled Feature Importance (DFI): Methodology

To overcome these limitations, the authors introduce Disentangled Feature Importance (DFI), a nonparametric generalization of the classical R2R^2 decomposition. The core idea is to transform the original feature space XX into a latent space Z=T(X)Z = T(X) where the coordinates are independent, using an optimal transport map TT. Feature importance is then computed in this disentangled space, free from the confounding effects of correlation, and attributed back to the original features via the sensitivity of the inverse map.

The DFI procedure consists of three stages:

  1. Disentanglement: Map XX to ZZ such that ZZ has independent coordinates (e.g., via the Bures-Wasserstein or Knothe-Rosenblatt transport).
  2. Latent Importance Computation: For each ZjZ_j, compute

ϕZj=E[Var(η(Z)Zj)]\phi_{Z_j} = \mathbb{E}\left[\mathrm{Var}(\eta(Z) \mid Z_{-j})\right]

where η(Z)=μ(T1(Z))\eta(Z) = \mu(T^{-1}(Z)).

  1. Attribution to Original Features: For each XlX_l, aggregate the latent importances weighted by the squared sensitivity of XlX_l to ZjZ_j:

ϕXl=j=1dE[Var(η(Z)Zj)(XlZj)2]\phi_{X_l} = \sum_{j=1}^d \mathbb{E}\left[\mathrm{Var}(\eta(Z) \mid Z_{-j}) \left(\frac{\partial X_l}{\partial Z_j}\right)^2\right]

This approach ensures that importance scores are both parsimonious (zero for truly irrelevant features) and exhaustive (nonzero for any feature that influences the outcome, directly or via dependencies).

Statistical Theory and Inference

The authors develop a comprehensive semiparametric theory for DFI. For general transport maps, they establish root-nn consistency and asymptotic normality of the importance estimators in the latent space, and extend these results to the original feature space for the Bures-Wasserstein map. The estimators achieve a second-order error rate, vanishing if both the regression function and transport map are estimated at rates faster than n1/4n^{-1/4}. The influence function for the attributed importance is derived explicitly, enabling valid confidence intervals and hypothesis testing.

A key computational advantage is that DFI avoids repeated submodel refitting (as in LOCO) and the need to estimate conditional covariate distributions (as in CPI), making it scalable to high-dimensional settings.

Empirical Evaluation

Simulation studies demonstrate that DFI consistently recovers the true importance structure, even under strong feature correlations and nonlinearities. In contrast, LOCO, CPI, and Shapley-based methods systematically underestimate importance in these regimes. DFI also provides accurate inferential coverage and is computationally efficient compared to decorrelated LOCO and Shapley sampling.

A real-data application to HIV-1 resistance prediction illustrates DFI's practical utility. The method identifies biologically meaningful groups of genomic features as most important, with statistical significance, and provides a nuanced ranking that aligns with domain knowledge.

Implications and Future Directions

Practical Implications:

  • DFI provides a principled, computationally efficient, and statistically robust framework for feature importance in the presence of arbitrary dependencies.
  • The method is directly applicable to high-dimensional, structured, and correlated data, such as genomics, NLP, and causal inference settings.
  • DFI's additive structure enables group-level importance analysis and is compatible with modern machine learning pipelines.

Theoretical Implications:

  • DFI generalizes the R2R^2 decomposition to nonparametric and nonlinear settings, preserving interpretability and variance attribution.
  • The framework connects feature importance to functional ANOVA and Sobol indices, providing a unified view of sensitivity analysis and model interpretability.

Limitations and Future Work:

  • The current theory assumes absolutely continuous covariate distributions; extensions to discrete or mixed-type data are needed.
  • Inference near the null remains challenging due to the quadratic nature of the functional; conservative interval expansion is suggested as a remedy.
  • Efficient estimation of high-dimensional transport maps is an open computational problem; leveraging generative models or scalable optimal transport solvers is a promising direction.
  • Extensions to non-squared-error losses, local (instance-level) importance, and incorporation of domain knowledge (e.g., feature hierarchies) are natural next steps.

Conclusion

Disentangled Feature Importance offers a rigorous and practical solution to the longstanding problem of feature attribution under correlation. By leveraging optimal transport and semiparametric theory, DFI enables reliable, interpretable, and computationally tractable importance estimation in complex, real-world data. The framework sets a new standard for feature importance analysis and opens avenues for further methodological and applied advances in interpretable machine learning.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 post and received 33 likes.

alphaXiv

  1. Disentangled Feature Importance (11 likes, 0 questions)