Disentangling Interactions and Dependencies in Feature Attribution (2410.23772v1)
Abstract: In explainable machine learning, global feature importance methods try to determine how much each individual feature contributes to predicting the target variable, resulting in one importance score for each feature. But often, predicting the target variable requires interactions between several features (such as in the XOR function), and features might have complex statistical dependencies that allow to partially replace one feature with another one. In commonly used feature importance scores these cooperative effects are conflated with the features' individual contributions, making them prone to misinterpretations. In this work, we derive DIP, a new mathematical decomposition of individual feature importance scores that disentangles three components: the standalone contribution and the contributions stemming from interactions and dependencies. We prove that the DIP decomposition is unique and show how it can be estimated in practice. Based on these results, we propose a new visualization of feature importance scores that clearly illustrates the different contributions.
- Barrett, A. B. (2015). Exploration of synergistic and redundant information sharing in static and dynamical gaussian systems. Physical Review E, 91(5):052802.
- From Shapley values to generalized additive models and back. In International Conference on Artificial Intelligence and Statistics (AISTATS).
- Breiman, L. (2001). Random forests. Machine learning, 45:5–32.
- Generalized sobol sensitivity indices for dependent variables: numerical methods. Journal of statistical computation and simulation, 85(7):1306–1333.
- True to the model or true to the data? arXiv preprint arXiv:2006.16234.
- Modeling wine preferences by data mining from physicochemical properties. Decision support systems, 47(4):547–553.
- Explaining by removing: A unified framework for model explanation. Journal of Machine Learning Research, 22(209):1–90.
- Understanding global feature contributions with additive importance measures. Neural Information Processing Systems (NeurIPS), 33.
- UCI machine learning repository.
- Scientific inference with interpretable machine learning: Analyzing models to learn about real-world phenomena. Minds and Machines, 34(3):32.
- Predictive learning via rule ensembles. The Annals of Applied Statistics, 2(3):916–954.
- Graphical views of suppression and multicollinearity in multiple linear regression. The American Statistician, 59(2):127–136.
- KernelSHAP-IQ: Weighted least-square optimization for Shapley interactions. arXiv preprint arXiv:2405.10852.
- SHAP-IQ: Unified approximation of any-order Shapley interactions. Neural Information Processing Systems (NeurIPS).
- Probabilistic sensitivity analysis with dependent variables: Covariance-based decomposition of hydrologic models. Water Resources Research, 59(4):e2022WR032834.
- Géron, A. (2022). Hands-on machine learning with Scikit-Learn, Keras, and TensorFlow. O’Reilly Media, Inc.
- An axiomatic approach to the concept of interaction among players in cooperative games. International Journal of game theory, 28:547–565.
- Quantifying redundant information in predicting a target random variable. Entropy, 17(7):4644–4653.
- Generalized additive models. Statist. Sci., 1(4):297–310.
- Decomposing global feature effects based on feature interactions. arXiv preprint arXiv:2306.00541.
- Statistical aspects of shap: Functional anova for model interpretation. arXiv preprint arXiv:2208.09970.
- Unifying local and global model explanations by functional decomposition of low dimensional structures. In International Conference on Artificial Intelligence and Statistics (AISTATS).
- Hooker, G. (2007). Generalized functional anova diagnostics for high-dimensional functions of dependent variables. Journal of computational and graphical statistics, 16(3):709–732.
- Unrestricted permutation forces extrapolation: variable importance requires at least one more model, or there is no free variable importance. Statistics and Computing, 31:1–16.
- Kolchinsky, A. (2022). A novel approach to the partial information decomposition. Entropy, 24(3):403.
- Relative feature importance. In International Conference on Pattern Recognition (ICPR).
- Distribution-free predictive inference for regression. Journal of the American Statistical Association, 113(523):1094–1111.
- Purifying interaction effects with the functional anova: An efficient algorithm for recovering identifiable additive models. In International Conference on Artificial Intelligence and Statistics (AISTATS).
- General formulation of hdmr component functions with independent and correlated variables. Journal of mathematical chemistry, 50:99–130.
- Luenberger, D. G. (1997). Optimization by vector space methods. John Wiley & Sons.
- Consistent individualized feature attribution for tree ensembles. arXiv preprint arXiv:1802.03888.
- A unified approach to interpreting model predictions. In Neural Information Processing Systems (NeurIPS).
- Molnar, C. (2022). Interpretable Machine Learning. 2 edition.
- Beyond treeSHAP: Efficient computation of any-order Shapley interactions for tree ensembles. In AAAI Conference on Artificial Intelligence.
- Interpreting multiple linear regression: a guidebook of variable importance. Practical assessment, research & evaluation, 17(9):n9.
- Interpretml: A unified framework for machine learning interpretability. arXiv preprint arXiv:1909.09223.
- Using commonality analysis in multiple regressions: a tool to decompose regression effects in the face of multicollinearity. Methods in Ecology and Evolution, 5(4):320–328.
- Commonality analysis: A method for decomposing explained variance in multiple regression analyses. Human Communication Research, 5(4):355–365.
- pygam: Generalized additive models in python. Zenodo.
- Shieh, G. (2006). Suppression situations in multiple linear regression. Educational and psychological measurement, 66(3):435–447.
- Shapley effects for global sensitivity analysis: Theory and computation. SIAM/ASA Journal on Uncertainty Quantification, 4(1):1060–1083.
- Conditional variable importance for random forests. BMC bioinformatics, 9(307):1–11.
- The Shapley Taylor interaction index. In International conference on machine learning (ICML).
- Nonnegative decomposition of multivariate information. arXiv preprint arXiv:1004.2515.
- Nonparametric variable importance assessment using machine learning techniques. Biometrics, 77(1):9–22.
- Withers, C. S. (1985). The moments of the multivariate normal. Bulletin of the Australian Mathematical Society, 32(1):103–107.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.