Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards trustable SHAP scores (2405.00076v2)

Published 30 Apr 2024 in cs.LG and cs.AI

Abstract: SHAP scores represent the proposed use of the well-known Shapley values in eXplainable Artificial Intelligence (XAI). Recent work has shown that the exact computation of SHAP scores can produce unsatisfactory results. Concretely, for some ML models, SHAP scores will mislead with respect to relative feature influence. To address these limitations, recently proposed alternatives exploit different axiomatic aggregations, all of which are defined in terms of abductive explanations. However, the proposed axiomatic aggregations are not Shapley values. This paper investigates how SHAP scores can be modified so as to extend axiomatic aggregations to the case of Shapley values in XAI. More importantly, the proposed new definition of SHAP scores avoids all the known cases where unsatisfactory results have been identified. The paper also characterizes the complexity of computing the novel definition of SHAP scores, highlighting families of classifiers for which computing these scores is tractable. Furthermore, the paper proposes modifications to the existing implementations of SHAP scores. These modifications eliminate some of the known limitations of SHAP scores, and have negligible impact in terms of performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (35)
  1. Towards rigorous interpretations: a formalisation of feature attribution. In ICML, pp.  76–86, 2021.
  2. The tractability of SHAP-score-based explanations for classification over deterministic and decomposable boolean circuits. In AAAI, pp.  6670–6678, 2021.
  3. On the complexity of SHAP-score-based explanations: Tractability via knowledge compilation and non-approximability results. J. Mach. Learn. Res., 24:63:1–63:58, 2023. URL http://jmlr.org/papers/v24/21-0389.html.
  4. Axiomatic aggregations of abductive explanations. CoRR, abs/2310.03131, 2023. doi: 10.48550/arXiv.2310.03131. URL https://doi.org/10.48550/arXiv.2310.03131.
  5. Exact Shapley values for local and model-true explanations of decision tree ensembles. Machine Learning with Applications, 9:100345, 2022.
  6. Darwiche, A. Logic for explainable AI. In LICS, pp.  1–11, 2023.
  7. A knowledge compilation map. J. Artif. Intell. Res., 17:229–264, 2002. doi: 10.1613/jair.989. URL https://doi.org/10.1613/jair.989.
  8. Dubey, P. On the uniqueness of the shapley value. International Journal of Game Theory, 4:131–139, 1975.
  9. Shapley values for feature selection: The good, the bad, and the axioms. IEEE Access, 9:144352–144360, 2021.
  10. The inadequacy of shapley values for explainability. CoRR, abs/2302.08160, 2023. doi: 10.48550/arXiv.2302.08160. URL https://doi.org/10.48550/arXiv.2302.08160.
  11. On the failings of shapley values for explainability. International Journal of Approximate Reasoning, pp.  109112, 2024. ISSN 0888-613X. doi: https://doi.org/10.1016/j.ijar.2023.109112. URL https://www.sciencedirect.com/science/article/pii/S0888613X23002438.
  12. Feature necessity & relevancy in ML classifier explanations. In TACAS, pp.  167–186, 2023.
  13. From contrastive to abductive explanations and back again. In AIxIA, pp.  335–355, 2020.
  14. Feature relevance quantification in explainable AI: A causal problem. In AISTATS, pp.  2907–2916, 2020.
  15. Shapley residuals: Quantifying the limits of the Shapley value for explanations. In NeurIPS, pp.  26598–26608, 2021.
  16. Problems with Shapley-value-based explanations as feature importance measures. In ICML, pp.  5491–5500, 2020.
  17. A unified approach to interpreting model predictions. In NeurIPS, pp.  4765–4774, 2017.
  18. Marques-Silva, J. Logic-based explainability in machine learning. In Reasoning Web, pp.  24–104, 2022.
  19. The explanation game: Explaining machine learning models using Shapley values. In CDMAKE, pp.  17–38, 2020.
  20. Towards unifying feature attribution and counterfactual explanations: Different means to the same end. In AIES, pp.  652–663, 2021.
  21. Roth, A. E. The Shapley value: essays in honor of Lloyd S. Shapley. Cambridge University Press, 1988.
  22. Shapley, L. S. A value for n𝑛nitalic_n-person games. Contributions to the Theory of Games, 2(28):307–317, 1953.
  23. An efficient explanation of individual classifications using game theory. J. Mach. Learn. Res., 11:1–18, 2010. URL https://dl.acm.org/doi/10.5555/1756006.1756007.
  24. Explaining prediction models and individual predictions with feature contributions. Knowl. Inf. Syst., 41(3):647–665, 2014. URL https://doi.org/10.1007/s10115-013-0679-x.
  25. The many shapley values for model explanation. In ICML, pp.  9269–9278, 2020.
  26. Valiant, L. G. The complexity of enumeration and reliability problems. SIAM J. Comput., 8(3):410–421, 1979. doi: 10.1137/0208032. URL https://doi.org/10.1137/0208032.
  27. On the tractability of SHAP explanations. In AAAI, pp.  6505–6513, 2021.
  28. On the tractability of SHAP explanations. J. Artif. Intell. Res., 74:851–886, 2022. doi: 10.1613/jair.1.13283. URL https://doi.org/10.1613/jair.1.13283.
  29. The computational complexity of understanding binary classifier decisions. J. Artif. Intell. Res., 70:351–387, 2021. doi: 10.1613/JAIR.1.12359. URL https://doi.org/10.1613/jair.1.12359.
  30. Local explanations via necessity and sufficiency: unifying theory and practice. In UAI, volume 161, pp.  1382–1392, 2021.
  31. If you like Shapley then you’ll love the core. In AAAI, pp.  5751–5759, 2021.
  32. Young, H. P. Monotonic solutions of cooperative games. International Journal of Game Theory, 14:65–72, 1985.
  33. Deep neural network or dermatologist? CoRR, abs/1908.06612, 2019. URL http://arxiv.org/abs/1908.06612.
  34. Anytime approximate formal feature attribution. CoRR, abs/2312.06973, 2023a. doi: 10.48550/ARXIV.2312.06973. URL https://doi.org/10.48550/arXiv.2312.06973.
  35. On formal feature attribution and its approximation. CoRR, abs/2307.03380, 2023b. doi: 10.48550/arXiv.2307.03380. URL https://doi.org/10.48550/arXiv.2307.03380.
Citations (2)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

Youtube Logo Streamline Icon: https://streamlinehq.com