Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

The Distributional Uncertainty of the SHAP score in Explainable Machine Learning (2401.12731v4)

Published 23 Jan 2024 in cs.AI, cs.LG, and cs.LO

Abstract: Attribution scores reflect how important the feature values in an input entity are for the output of a machine learning model. One of the most popular attribution scores is the SHAP score, which is an instantiation of the general Shapley value used in coalition game theory. The definition of this score relies on a probability distribution on the entity population. Since the exact distribution is generally unknown, it needs to be assigned subjectively or be estimated from data, which may lead to misleading feature scores. In this paper, we propose a principled framework for reasoning on SHAP scores under unknown entity population distributions. In our framework, we consider an uncertainty region that contains the potential distributions, and the SHAP score of a feature becomes a function defined over this region. We study the basic problems of finding maxima and minima of this function, which allows us to determine tight ranges for the SHAP scores of all features. In particular, we pinpoint the complexity of these problems, and other related ones, showing them to be NP-complete. Finally, we present experiments on a real-world dataset, showing that our framework may contribute to a more robust feature scoring.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (23)
  1. D Alvarez-Melis and T Jaakkola. On the Robustness of Interpretability Methods. ArXiv preprint: 1806.08049, 2018.
  2. Scalar Aggregation in Inconsistent Databases. Theoretical Computer Science, 296(3):405–434, 2003.
  3. On the Complexity of SHAP-Score- Based Explanations: Tractability via Knowledge Compilation and Non-Approximability Results. J. Machine Learning Research, 24(63):1–58, 2023.
  4. Model Interpretability through the Lens of Computational Complexity. Advances in Neural Information Processing Systems, 2020.
  5. L Bertossi and J E León. Efficient Computation of Shap Explanation Scores for Neural Network Classifiers via Knowledge Compilation. In European Conference on Logics in Artificial Intelligence, pages 49–64. Springer, 2023.
  6. Causality-based Explanation of Classification Outcomes. Proc. 4th SIGMOD Int. WS on Data Management for End-to-End Machine Learning, pages 70–81, 2020.
  7. Machine Learning Interpretability: A Survey on Methods and Metrics. Electronics, 8(8):832, 2019.
  8. A Darwiche and P Marquis. A Knowledge Compilation Map. J. Artif. Int. Res., 17(1):229–264, 2002.
  9. A Darwiche. On the Tractable Counting of Theory Models and its Application to Truth Maintenance and Belief Revision. Journal of Applied Non-Classical Logics, 11(1-2):11–34, 2001.
  10. X Huang and J Marques-Silva. From Robustness to Explainability and Back Again. ArXiv preprint: 2306.03048, 2023.
  11. The Interval Analysis of Multilinear Expressions. Electronic Notes in Theoretical Computer Science, 267(2):43–53, 2010.
  12. Explainable AI: A Review of Machine Learning Interpretability Methods. Entropy, 23(1):18, 2020.
  13. A Unified Approach to Interpreting Model Predictions. Advances in Neural Information Processing Systems, 30, 2017.
  14. From Local Explanations to Global Understanding with Explainable AI for Trees. Nature Machine Intelligence, 2(1):2522–5839, 2020.
  15. Christoph Molnar. Interpretable Machine Learning. 2020.
  16. Nugent, C. California Housing Prices. https://www.kaggle.com/datasets/camnugent/california-housing-prices, 2018.
  17. ” why should i trust you?” explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pages 1135–1144, 2016.
  18. A E Roth, editor. The Shapley Value: Essays in Honor of Lloyd S. Shapley. Cambridge Univ. Press, 1988.
  19. L. S. Shapley. A Value for n-Person Games. Contributions to the Theory of Games, 2(28):307–317, 1953.
  20. Learning Important Features through Propagating Activation Differences. In International Conference on Machine Learning, pages 3145–3153. PMLR, 2017.
  21. Reliable Post Hoc Explanations: Modeling Uncertainty in Explainability. Advances in Neural Information Processing Systems, 34:9391–9404, 2021.
  22. E Štrumbelj and I Kononenko. Explaining Prediction Models and Individual Predictions with Feature Contributions. Knowledge and Information Systems, 41:647–665, 2014.
  23. On the Tractability of SHAP Explanations. J. Artif. Intell. Res., 74:851–886, 2022.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets