Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Is Ignorance Bliss? The Role of Post Hoc Explanation Faithfulness and Alignment in Model Trust in Laypeople and Domain Experts (2312.05690v2)

Published 9 Dec 2023 in cs.HC

Abstract: Post hoc explanations have emerged as a way to improve user trust in machine learning models by providing insight into model decision-making. However, explanations tend to be evaluated based on their alignment with prior knowledge while the faithfulness of an explanation with respect to the model, a fundamental criterion, is often overlooked. Furthermore, the effect of explanation faithfulness and alignment on user trust and whether this effect differs among laypeople and domain experts is unclear. To investigate these questions, we conduct a user study with computer science students and doctors in three domain areas, controlling the laypeople and domain expert groups in each setting. The results indicate that laypeople base their trust in explanations on explanation faithfulness while domain experts base theirs on explanation alignment. To our knowledge, this work is the first to show that (1) different factors affect laypeople and domain experts' trust in post hoc explanations and (2) domain experts are subject to specific biases due to their expertise when interpreting post hoc explanations. By uncovering this phenomenon and exposing this cognitive bias, this work motivates the need to educate end users about how to properly interpret explanations and overcome their own cognitive biases, and motivates the development of simple and interpretable faithfulness metrics for end users. This research is particularly important and timely as post hoc explanations are increasingly being used in high-stakes, real-world settings such as medicine.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (16)
  1. Sanity checks for saliency maps. Neural Information Processing Systems, 2018.
  2. Alejandro Barredo Arrieta et al. Explainable artificial intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion, 2020.
  3. The role of domain expertise in trusting and following explainable AI decision support systems. Journal of Decision Systems, 2021.
  4. Longbing Cao. AI in finance: Challenges, techniques, and opportunities. ACM Computing Surveys (CSUR), 2022.
  5. The effects of domain knowledge on trust in explainable AI and task performance: A case of peer-to-peer lending. International Journal of Human-Computer Studies, 2022.
  6. A comparison of explanations given by explainable artificial intelligence methods on analysing electronic health records. IEEE EMBS International Conference on Biomedical and Health Informatics, 2021.
  7. What are people doing about XAI user experience? A survey on AI explainability research and practice. Design, User Experience, and Usability. Design for Contemporary Interactive Environments. HCI International Conference, 2020.
  8. Which explanation should I choose? A function approximation perspective to characterizing post hoc explanations. Neural Information Processing Systems, 2022.
  9. A benchmark for interpretability methods in deep neural networks. Neural Information Processing Systems, 2019.
  10. “Why should I trust you?” Explaining the predictions of any classifier. ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2016.
  11. Benchmarking saliency methods for chest x-ray interpretation. Nature Machine Intelligence, 2022.
  12. SmoothGrad: Removing noise by adding noise. arXiv preprint arXiv:1706.03825, 2017.
  13. Axiomatic attribution for deep networks. International Conference on Machine Learning, 2017.
  14. Cyber Security, Artificial Intelligence, Data Protection and the Law. Springer, 2021.
  15. Are explanations helpful? A comparative study of the effects of explanations in AI-assisted decision-making. International Conference on Intelligent User Interfaces, 2021.
  16. Artificial intelligence in healthcare. Nature Biomedical Engineering, 2018.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Tessa Han (7 papers)
  2. Yasha Ektefaie (5 papers)
  3. Maha Farhat (2 papers)
  4. Marinka Zitnik (79 papers)
  5. Himabindu Lakkaraju (88 papers)
Citations (1)