Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
173 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Axiomatic Aggregations of Abductive Explanations (2310.03131v3)

Published 29 Sep 2023 in cs.AI

Abstract: The recent criticisms of the robustness of post hoc model approximation explanation methods (like LIME and SHAP) have led to the rise of model-precise abductive explanations. For each data point, abductive explanations provide a minimal subset of features that are sufficient to generate the outcome. While theoretically sound and rigorous, abductive explanations suffer from a major issue -- there can be several valid abductive explanations for the same data point. In such cases, providing a single abductive explanation can be insufficient; on the other hand, providing all valid abductive explanations can be incomprehensible due to their size. In this work, we solve this issue by aggregating the many possible abductive explanations into feature importance scores. We propose three aggregation methods: two based on power indices from cooperative game theory and a third based on a well-known measure of causal strength. We characterize these three methods axiomatically, showing that each of them uniquely satisfies a set of desirable properties. We also evaluate them on multiple datasets and show that these explanations are robust to the attacks that fool SHAP and LIME.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. Machine bias: There’s software used across the country to predict future criminals. and it’s biased against blacks. ProPublica, May 2016.
  2. On computing probabilistic explanations for decision trees. In Proceedings of the 35th Annual Conference on Neural Information Processing Systems (NeurIPS), 2022.
  3. On preferred abductive explanations for decision trees and random forests. In Proceedings of the 31st International Joint Conference on Artificial Intelligence (IJCAI), pages 643–650, 2022.
  4. The hidden assumptions behind counterfactual explanations and principal reasons. In Proceedings of the 3rd ACM Conference on Fairness, Accountability, and Transparency (FAccT), pages 80–89, 2020.
  5. Computational Aspects of Cooperative Game Theory. Morgan & Claypool Publishers, 1st edition, 2011.
  6. Responsibility and blame: A structural-model approach. Journal of Artificial Intelligence Research, 22:93–115, 2004.
  7. Algorithmic transparency via quantitative input influence: Theory and experiments with learning systems. IEEE Symposium on Security and Privacy, pages 598–617, 2016.
  8. J. Deegan and Edward Packel. A new index of power for simple n-person games. International Journal of Game Theory, 7:113–123, 1978.
  9. UCI machine learning repository, 2017. URL http://archive.ics.uci.edu/ml.
  10. Asymmetric shapley values: Incorporating causal knowledge into model-agnostic explainability. In Proceedings of the 34th Annual Conference on Neural Information Processing Systems (NeurIPS), 2020.
  11. Explaining black-box algorithms using probabilistic contrastive counterfactuals, 2021.
  12. Manfred J. Holler. Forming coalitions and measuring voting power. Political Studies, 30:262–271, 1982.
  13. Power, luck and the right index. Journal of Economics, 43:21–29, 1983.
  14. The inadequacy of shapley values for explainability, 2023.
  15. Abduction-based explanations for machine learning models. In Proceedings of the 33rd AAAI Conference on Artificial Intelligence (AAAI), 2019a.
  16. On validating, repairing and refining heuristic ML explanations. CoRR, abs/1907.02509, 2019b.
  17. From contrastive to abductive explanations and back again. In Matteo Baldoni and Stefania Bandini, editors, Proceedings of the 19thInternational Conference of the Italian Association for Artificial Intelligence (AIxIA), volume 12414, pages 335–355. Springer, 2020a.
  18. On formal reasoning about explanations. In RCRA, 2020b.
  19. Using maxsat for efficient explanations of tree ensembles. In Proceedings of the 36th AAAI Conference on Artificial Intelligence (AAAI), pages 3776–3785, 2022.
  20. On computing probabilistic abductive explanations. Int. J. Approx. Reason., 159, 2023.
  21. Fast, flexible MUS enumeration. Constraints An Int. J., 21(2):223–250, 2016.
  22. A unified approach to interpreting model predictions. In Proceedings of the 31st Annual Conference on Neural Information Processing Systems (NeurIPS), pages 4768–4777, 2017.
  23. Explaining naive bayes and other linear classifiers with polynomial time and delay. In Proceedings of the 34th Annual Conference on Neural Information Processing Systems (NeurIPS), 2020.
  24. Explanations for monotonic classifiers. In Proceedings of the 38th International Conference on Machine Learning (ICML), pages 7469–7479, 2021.
  25. Deep learning for financial applications : A survey. Applied Soft Computing, 93:106384, 2020. ISSN 1568-4946.
  26. A comprehensive survey of deep learning in the field of medical imaging and medical natural language processing: Challenges and research directions. Journal of King Saud University - Computer and Information Sciences, 34:5083–5099, 2022. ISSN 1319-1578.
  27. High dimensional model explanations: An axiomatic approach. In Proceedings of the 4th ACM Conference on Fairness, Accountability, and Transparency (FAccT), pages 401–411. ACM, 2021.
  28. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830, 2011.
  29. Secure and robust machine learning for healthcare: A survey. IEEE Reviews in Biomedical Engineering, 14:156–180, 2021.
  30. ”why should i trust you?”: Explaining the predictions of any classifier. In Proceedings of the 22nd International Conference on Knowledge Discovery and Data Mining (KDD), page 1135–1144, 2016.
  31. Cynthia Rudin. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature machine intelligence, 1(5):206–215, 2019.
  32. Lloyd S. Shapley. A Value for n-Person Games, pages 307–318. Princeton University Press, 1953.
  33. A symbolic approach to explaining bayesian network classifiers. In Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI), page 5103–5111, 2018.
  34. Fooling lime and shap: Adversarial attacks on post hoc explanation methods. In Proceedings of the 3rd AAAI/ACM Conference on Artifical Intelligence, Ethics, and Society (AIES), 2020.
  35. The many shapley values for model explanation. In Proceedings of the 37th International Conference on Machine Learning (ICML), pages 9269–9278, 2020.
  36. Axiomatic attribution for deep networks. In Proceedings of the 34th International Conference on Machine Learning (ICML), page 3319–3328, 2017.
  37. The computational complexity of understanding binary classifier decisions. J. Artif. Intell. Res., 70:351–387, 2021.
  38. Local explanations via necessity and sufficiency: Unifying theory and practice, 2021.
  39. H. Young. Monotonic solutions of cooperative games. International Journal of Game Theory, 14:65–72, 1985.
  40. On formal feature attribution and its approximation. CoRR, abs/2307.03380, 2023.
Citations (2)

Summary

We haven't generated a summary for this paper yet.