Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

QUCE: The Minimisation and Quantification of Path-Based Uncertainty for Generative Counterfactual Explanations (2402.17516v3)

Published 27 Feb 2024 in cs.LG and cs.AI

Abstract: Deep Neural Networks (DNNs) stand out as one of the most prominent approaches within the Machine Learning (ML) domain. The efficacy of DNNs has surged alongside recent increases in computational capacity, allowing these approaches to scale to significant complexities for addressing predictive challenges in big data. However, as the complexity of DNN models rises, interpretability diminishes. In response to this challenge, explainable models such as Adversarial Gradient Integration (AGI) leverage path-based gradients provided by DNNs to elucidate their decisions. Yet the performance of path-based explainers can be compromised when gradients exhibit irregularities during out-of-distribution path traversal. In this context, we introduce Quantified Uncertainty Counterfactual Explanations (QUCE), a method designed to mitigate out-of-distribution traversal by minimizing path uncertainty. QUCE not only quantifies uncertainty when presenting explanations but also generates more certain counterfactual examples. We showcase the performance of the QUCE method by comparing it with competing methods for both path-based explanations and generative counterfactual examples.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (35)
  1. Towards credible visual model interpretation with path attribution. In Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pages 439–457. PMLR, 23–29 Jul 2023. URL https://proceedings.mlr.press/v202/akhtar23a.html.
  2. L. Amgoud and J. Ben-Naim. Axiomatic foundations of explainability. In IJCAI, pages 636–642, 2022.
  3. Reconstruction error-based anomaly detection with few outlying examples, 2023.
  4. Getting a CLUE: A method for explaining uncertainty estimates. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=XSLF1XFq5h.
  5. Selecting the most appropriate treatment for each patient. International Journal of Clinical and Health Psychology, 16(1):99–108, Jan. 2016. ISSN 1697-2600. 10.1016/j.ijchp.2015.08.001. URL http://dx.doi.org/10.1016/j.ijchp.2015.08.001.
  6. A review on explainable artificial intelligence for healthcare: Why, how, and when? IEEE Transactions on Artificial Intelligence, pages 1–15, 2023. 10.1109/TAI.2023.3266418.
  7. Reconciling multiple counterfactuals when evaluating biodiversity conservation impact in social-ecological systems. Conservation Biology, 35(2):510–521, Sept. 2020. ISSN 1523-1739. 10.1111/cobi.13570. URL http://dx.doi.org/10.1111/cobi.13570.
  8. Explainable artificial intelligence in finance: A bibliometric review. Finance Research Letters, 56:104145, 2023. ISSN 1544-6123. https://doi.org/10.1016/j.frl.2023.104145. URL https://www.sciencedirect.com/science/article/pii/S1544612323005172.
  9. Cafe: Conflict-aware feature-wise explanations, 2023.
  10. Batch integrated gradients: Explanations for temporal electronic health records. In Artificial Intelligence in Medicine, pages 120–124, Cham, 2023a. Springer Nature Switzerland. ISBN 978-3-031-34344-5.
  11. A formal introduction to batch-integrated gradients for temporal explanations. In 2023 IEEE 35th International Conference on Tools with Artificial Intelligence (ICTAI), pages 452–459, Los Alamitos, CA, USA, nov 2023b. IEEE Computer Society. 10.1109/ICTAI59109.2023.00072. URL https://doi.ieeecomputersociety.org/10.1109/ICTAI59109.2023.00072.
  12. Improving performance of deep learning models with axiomatic attribution priors and expected gradients. Nature Machine Intelligence, 3(7):620–631, May 2021. ISSN 2522-5839. 10.1038/s42256-021-00343-w. URL http://dx.doi.org/10.1038/s42256-021-00343-w.
  13. R. Guidotti. Counterfactual explanations and how to find them: literature review and benchmarking. Data Mining and Knowledge Discovery, Apr. 2022. ISSN 1573-756X. 10.1007/s10618-022-00831-6. URL http://dx.doi.org/10.1007/s10618-022-00831-6.
  14. Post-hoc counterfactual generation with supervised autoencoder. In Machine Learning and Principles and Practice of Knowledge Discovery in Databases, pages 105–114, Cham, 2021. Springer International Publishing. ISBN 978-3-030-93736-2.
  15. M. Höfler. Causal inference based on counterfactuals. BMC Medical Research Methodology, 5(1), Sept. 2005. ISSN 1471-2288. 10.1186/1471-2288-5-28. URL http://dx.doi.org/10.1186/1471-2288-5-28.
  16. Exmed: An ai tool for experimenting explainable ai techniques on medical data analytics. In 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pages 841–845, 2021. 10.1109/ICTAI52525.2021.00134.
  17. Guided integrated gradients: an adaptive path method for removing noise. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 5048–5056, Los Alamitos, CA, USA, jun 2021. IEEE Computer Society. 10.1109/CVPR46437.2021.00501. URL https://doi.ieeecomputersociety.org/10.1109/CVPR46437.2021.00501.
  18. On understanding the influence of controllable factors with a feature attribution algorithm: a medical case study. In INISTA, pages 1–8, 2022.
  19. Negative flux aggregation to estimate feature attributions. In E. Elkind, editor, Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI-23, pages 446–454. International Joint Conferences on Artificial Intelligence Organization, 8 2023. 10.24963/ijcai.2023/50. URL https://doi.org/10.24963/ijcai.2023/50. Main Track.
  20. S. M. Lundberg and S. Lee. A unified approach to interpreting model predictions. In NeurIPS, pages 4765–4774, 2017.
  21. Clear: Generative counterfactual explanations on graphs. In S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, editors, Advances in Neural Information Processing Systems, volume 35, pages 25895–25907. Curran Associates, Inc., 2022. URL https://proceedings.neurips.cc/paper_files/paper/2022/file/a69d7f3a1340d55c720e572742439eaf-Paper-Conference.pdf.
  22. Uncertainty in additive feature attribution methods, 2023.
  23. Explaining machine learning classifiers through diverse counterfactual explanations. In FAccT, page 607–617, 2020.
  24. Explaining deep neural network models with adversarial gradient integration. In Z.-H. Zhou, editor, Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21, pages 2876–2883. International Joint Conferences on Artificial Intelligence Organization, 8 2021. 10.24963/ijcai.2021/396. URL https://doi.org/10.24963/ijcai.2021/396. Main Track.
  25. Causal inference and counterfactual prediction in machine learning for actionable healthcare. Nature Machine Intelligence, 2(7):369–375, July 2020. ISSN 2522-5839. 10.1038/s42256-020-0197-y. URL http://dx.doi.org/10.1038/s42256-020-0197-y.
  26. "Why should I trust you?": Explaining the predictions of any classifier. In ACM SIGKDD, page 1135–1144, 2016.
  27. A. Sagar. Uncertainty quantification using variational inference for biomedical image segmentation. In 2022 IEEE/CVF Winter Conference on Applications of Computer Vision Workshops (WACVW), pages 44–51, Los Alamitos, CA, USA, jan 2022. IEEE Computer Society. 10.1109/WACVW54805.2022.00010. URL https://doi.ieeecomputersociety.org/10.1109/WACVW54805.2022.00010.
  28. A stochastic approximation method. IEEE Transactions on Systems, Man, and Cybernetics, SMC-1(4):338–344, 1971. 10.1109/TSMC.1971.4308316.
  29. Reliable post hoc explanations: Modeling uncertainty in explainability. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P. Liang, and J. W. Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 9391–9404. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper_files/paper/2021/file/4e246a381baf2ce038b3b0f82c7d6fb4-Paper.pdf.
  30. Axiomatic attribution for deep networks. page 3319–3328, 2017.
  31. Practical autoencoder based anomaly detection by using vector reconstruction error. Cybersecurity, 6(1), Jan. 2023. ISSN 2523-3246. 10.1186/s42400-022-00134-9. URL http://dx.doi.org/10.1186/s42400-022-00134-9.
  32. "counterfactual explanations without opening the black box: Automated decisions and the gdpr. In arXiv, 2017.
  33. Breast Cancer Wisconsin (Diagnostic). UCI Machine Learning Repository, 1995. DOI: https://doi.org/10.24432/C5DW2B.
  34. Local path integration for attribution. Proceedings of the AAAI Conference on Artificial Intelligence, 37(3):3173–3180, Jun. 2023a. 10.1609/aaai.v37i3.25422. URL https://ojs.aaai.org/index.php/AAAI/article/view/25422.
  35. IDGI: A framework to eliminate explanation noise from integrated gradients. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2023b. URL http://www.cs.iit.edu/~ml/pdfs/yang-cvpr23.pdf.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Jamie Duell (2 papers)
  2. Hsuan Fu (3 papers)
  3. Monika Seisenberger (6 papers)
  4. Xiuyi Fan (13 papers)

Summary

We haven't generated a summary for this paper yet.