Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 72 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 30 tok/s Pro
GPT-5 High 24 tok/s Pro
GPT-4o 115 tok/s Pro
Kimi K2 203 tok/s Pro
GPT OSS 120B 451 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

Disentangling Interactions and Dependencies in Feature Attribution (2410.23772v1)

Published 31 Oct 2024 in cs.LG and stat.ML

Abstract: In explainable machine learning, global feature importance methods try to determine how much each individual feature contributes to predicting the target variable, resulting in one importance score for each feature. But often, predicting the target variable requires interactions between several features (such as in the XOR function), and features might have complex statistical dependencies that allow to partially replace one feature with another one. In commonly used feature importance scores these cooperative effects are conflated with the features' individual contributions, making them prone to misinterpretations. In this work, we derive DIP, a new mathematical decomposition of individual feature importance scores that disentangles three components: the standalone contribution and the contributions stemming from interactions and dependencies. We prove that the DIP decomposition is unique and show how it can be estimated in practice. Based on these results, we propose a new visualization of feature importance scores that clearly illustrates the different contributions.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (46)
  1. Barrett, A. B. (2015). Exploration of synergistic and redundant information sharing in static and dynamical gaussian systems. Physical Review E, 91(5):052802.
  2. From Shapley values to generalized additive models and back. In International Conference on Artificial Intelligence and Statistics (AISTATS).
  3. Breiman, L. (2001). Random forests. Machine learning, 45:5–32.
  4. Generalized sobol sensitivity indices for dependent variables: numerical methods. Journal of statistical computation and simulation, 85(7):1306–1333.
  5. True to the model or true to the data? arXiv preprint arXiv:2006.16234.
  6. Modeling wine preferences by data mining from physicochemical properties. Decision support systems, 47(4):547–553.
  7. Explaining by removing: A unified framework for model explanation. Journal of Machine Learning Research, 22(209):1–90.
  8. Understanding global feature contributions with additive importance measures. Neural Information Processing Systems (NeurIPS), 33.
  9. UCI machine learning repository.
  10. Scientific inference with interpretable machine learning: Analyzing models to learn about real-world phenomena. Minds and Machines, 34(3):32.
  11. Predictive learning via rule ensembles. The Annals of Applied Statistics, 2(3):916–954.
  12. Graphical views of suppression and multicollinearity in multiple linear regression. The American Statistician, 59(2):127–136.
  13. KernelSHAP-IQ: Weighted least-square optimization for Shapley interactions. arXiv preprint arXiv:2405.10852.
  14. SHAP-IQ: Unified approximation of any-order Shapley interactions. Neural Information Processing Systems (NeurIPS).
  15. Probabilistic sensitivity analysis with dependent variables: Covariance-based decomposition of hydrologic models. Water Resources Research, 59(4):e2022WR032834.
  16. Géron, A. (2022). Hands-on machine learning with Scikit-Learn, Keras, and TensorFlow. O’Reilly Media, Inc.
  17. An axiomatic approach to the concept of interaction among players in cooperative games. International Journal of game theory, 28:547–565.
  18. Quantifying redundant information in predicting a target random variable. Entropy, 17(7):4644–4653.
  19. Generalized additive models. Statist. Sci., 1(4):297–310.
  20. Decomposing global feature effects based on feature interactions. arXiv preprint arXiv:2306.00541.
  21. Statistical aspects of shap: Functional anova for model interpretation. arXiv preprint arXiv:2208.09970.
  22. Unifying local and global model explanations by functional decomposition of low dimensional structures. In International Conference on Artificial Intelligence and Statistics (AISTATS).
  23. Hooker, G. (2007). Generalized functional anova diagnostics for high-dimensional functions of dependent variables. Journal of computational and graphical statistics, 16(3):709–732.
  24. Unrestricted permutation forces extrapolation: variable importance requires at least one more model, or there is no free variable importance. Statistics and Computing, 31:1–16.
  25. Kolchinsky, A. (2022). A novel approach to the partial information decomposition. Entropy, 24(3):403.
  26. Relative feature importance. In International Conference on Pattern Recognition (ICPR).
  27. Distribution-free predictive inference for regression. Journal of the American Statistical Association, 113(523):1094–1111.
  28. Purifying interaction effects with the functional anova: An efficient algorithm for recovering identifiable additive models. In International Conference on Artificial Intelligence and Statistics (AISTATS).
  29. General formulation of hdmr component functions with independent and correlated variables. Journal of mathematical chemistry, 50:99–130.
  30. Luenberger, D. G. (1997). Optimization by vector space methods. John Wiley & Sons.
  31. Consistent individualized feature attribution for tree ensembles. arXiv preprint arXiv:1802.03888.
  32. A unified approach to interpreting model predictions. In Neural Information Processing Systems (NeurIPS).
  33. Molnar, C. (2022). Interpretable Machine Learning. 2 edition.
  34. Beyond treeSHAP: Efficient computation of any-order Shapley interactions for tree ensembles. In AAAI Conference on Artificial Intelligence.
  35. Interpreting multiple linear regression: a guidebook of variable importance. Practical assessment, research & evaluation, 17(9):n9.
  36. Interpretml: A unified framework for machine learning interpretability. arXiv preprint arXiv:1909.09223.
  37. Using commonality analysis in multiple regressions: a tool to decompose regression effects in the face of multicollinearity. Methods in Ecology and Evolution, 5(4):320–328.
  38. Commonality analysis: A method for decomposing explained variance in multiple regression analyses. Human Communication Research, 5(4):355–365.
  39. pygam: Generalized additive models in python. Zenodo.
  40. Shieh, G. (2006). Suppression situations in multiple linear regression. Educational and psychological measurement, 66(3):435–447.
  41. Shapley effects for global sensitivity analysis: Theory and computation. SIAM/ASA Journal on Uncertainty Quantification, 4(1):1060–1083.
  42. Conditional variable importance for random forests. BMC bioinformatics, 9(307):1–11.
  43. The Shapley Taylor interaction index. In International conference on machine learning (ICML).
  44. Nonnegative decomposition of multivariate information. arXiv preprint arXiv:1004.2515.
  45. Nonparametric variable importance assessment using machine learning techniques. Biometrics, 77(1):9–22.
  46. Withers, C. S. (1985). The moments of the multivariate normal. Bulletin of the Australian Mathematical Society, 32(1):103–107.

Summary

  • The paper introduces DIP, a novel method that decomposes feature importance into standalone, interaction, and dependency contributions.
  • It employs a Generalized Groupwise Additive Model to isolate pure interaction effects from main effects for clearer model diagnostics.
  • Evaluations on real datasets reveal that many features derive their predictive power mainly through interactions rather than standalone effects.

Disentangling Interactions and Dependencies in Feature Attribution

The paper "Disentangling Interactions and Dependencies in Feature Attribution" tackles a nuanced problem in the domain of explainable machine learning: the accurate assessment of feature importance when interactions and dependencies among features are present. This issue is vital as conventional feature attribution methods often conflate these aspects, leading to misinterpretations of feature contributions.

Key Contributions

The authors introduce a novel method named DIP (Disentangling Interactions and Dependencies) which decomposes feature importance into three distinct components: standalone contributions of the features, contributions from their interactions, and contributions from dependencies among them. The theoretical underpinning of this method is robust, backed by a proof that demonstrates its uniqueness in decomposition.

Approach and Methodology

At the core of this paper is the differentiation between a feature’s standalone effect and the cooperative effects it might exhibit in conjunction with other features. Traditional global feature importance methodologies mostly offer a single score per feature, which does not account for these intricacies. The framework introduced, DIP, addresses this shortfall by clearly visualizing the different contribution components, thereby providing a transparent estimation of feature value in a model's predictive capacity.

To implement DIP, the authors utilize an estimation procedure that involves fitting a Generalized Groupwise Additive Model (GGAM) to account for main effects while isolating pure interactions. This approach is significant because it allows the decomposition to separate pure interaction effects from main effect contributions by assessing how much information about the target can only be accessed through between-group cooperations.

Practical Implementations and Results

The authors evaluate DIP on real-world datasets, such as the wine quality and California housing datasets, where algorithmic performance metrics like Leave-One-Covariate-Out (LOCO) scores were calculated and analyzed. The findings indicate that many variables, previously thought individually important due to LOCO scores, actually derive their significance from interactions with other features. For example, longitude and latitude variables in housing data showed high importance scores primarily due to their interactions, not their standalone predictive power.

Furthermore, the authors demonstrate that many perceived standalone important features are actually due to redundant contributions related to dependencies with other features. This enhanced understanding pushed forward by DIP could lead to improved model diagnostics and better feature engineering practices by accurately delineating the role of feature interactions.

Theoretical Implications and Future Developments

The DIP methodology offers a fresh perspective that extends beyond classical interpretations of interaction effects in predictive modeling. The approach is primarily built on L2L^2 loss metrics but has potential applicability within other loss paradigms, such as cross-entropy losses widely used in binary classification tasks.

The resolution of interpreting feature importance through this lens suggests several avenues for future research. One direction is adapting this decomposition procedure for more model-specific explainer methods, which could yield an even more refined understanding of model behavior, especially in deep learning contexts where interactions are inherently more complex.

Another frontier involves tackling the computational aspects of feature attribution. While the complexity introduced by pairwise consideration is manageable as presented, the implications of extending this to larger feature sets and ensuring computational efficiency remain areas ripe for exploration. Furthermore, exploring machine learning models' fairness through this decomposition could uncover subtle biases encoded via interactions and dependencies.

Conclusion

In conclusion, the DIP framework represents a vital step forward in the interpretation of machine learning models. By disentangling interaction effects and dependencies, the authors enrich the landscape of explainable artificial intelligence, offering researchers and practitioners a powerful tool for better understanding the intricate details of feature interactions and their impact on predictive modeling. This approach not only enhances transparency but also paves the way for more reliable model diagnostics and effective policy and decision-making informed by machine learning insights.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 post and received 5 likes.