Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

From Feature Importance to Natural Language Explanations Using LLMs with RAG (2407.20990v1)

Published 30 Jul 2024 in cs.AI, cs.CL, cs.CV, cs.HC, and cs.LG
From Feature Importance to Natural Language Explanations Using LLMs with RAG

Abstract: As machine learning becomes increasingly integral to autonomous decision-making processes involving human interaction, the necessity of comprehending the model's outputs through conversational means increases. Most recently, foundation models are being explored for their potential as post hoc explainers, providing a pathway to elucidate the decision-making mechanisms of predictive models. In this work, we introduce traceable question-answering, leveraging an external knowledge repository to inform the responses of LLMs to user queries within a scene understanding task. This knowledge repository comprises contextual details regarding the model's output, containing high-level features, feature importance, and alternative probabilities. We employ subtractive counterfactual reasoning to compute feature importance, a method that entails analysing output variations resulting from decomposing semantic features. Furthermore, to maintain a seamless conversational flow, we integrate four key characteristics - social, causal, selective, and contrastive - drawn from social science research on human explanations into a single-shot prompt, guiding the response generation process. Our evaluation demonstrates that explanations generated by the LLMs encompassed these elements, indicating its potential to bridge the gap between complex model outputs and natural language expressions.

From Feature Importance to Natural Language Explanations Using LLMs with RAG

This paper addresses the complexities involved in making AI systems interpretable in a way that is accessible to non-technical users. It proposes a novel approach to explainable AI (XAI) by integrating LLMs with a decomposition-based methodology for extracting feature importance. The focus is on generating natural language explanations for scene-understanding tasks, which are often crucial in autonomous systems, including autonomous vehicles.

The core contribution is a method referred to as traceable question-answering through Retrieval-Augmented Generation (RAG). This technique enhances LLMs by referencing an external knowledge base, thereby resolving hallucination issues and ensuring that responses are grounded in verifiable model outputs. The external repository contains contextual data such as feature importance and predictions, enabling LLMs to act as reasoning engines rather than sources of information.

Methodology

The paper introduces a decomposition-based approach to infer feature importance, which involves systematically removing input components and observing changes in model outputs. This approach produces feature importance values that help in understanding model behavior. These values are then utilized in conjunction with a system prompt designed to imbue the LLMs with characteristics derived from social science research on human explanations. These characteristics include being social, causal, selective, and contrastive.

To achieve this, the system prompt directs the LLMs to format responses in a manner that mirrors human conversational norms, leveraging politeness and empathy while maintaining technical accuracy. The methodology is demonstrated through an application in scene-understanding tasks, specifically using urban scene segmentation data.

Evaluation and Results

Evaluation of the proposed approach is conducted using two models: GPT-3.5 and GPT-4. Their ability to generate explanations is assessed across sociability, causality, selectiveness, and contrastiveness. Results indicate that GPT-4 outperforms GPT-3.5 in all categories, showcasing higher accuracy and better causal inference capabilities. GPT-4's responses exhibit more nuanced expressions of causality and contrastiveness, confirming its potential for generating effective explanations.

The paper also evaluates simplicity in explanations, accounting for the frequency of technical jargon, response length, and the number of causes cited. GPT-4 again stands out by providing simpler and more comprehensible responses, suggesting its effectiveness in bridging the gap between complex model outputs and user understanding.

Implications and Future Directions

The research has important implications for advancing XAI, emphasizing the role of human-friendly explanations in enhancing trust and usability of AI systems. By integrating semantic feature importance with LLMs, the approach offers a potentially impactful solution for generating clarity in AI-driven decision processes. The dynamic nature of the external knowledge base further adds to the adaptability and applicability of this method across various domains.

Future research directions could explore enhanced dialogue management strategies to improve interaction personalization, potentially integrating multimodal elements for richer explanations. Additionally, cross-algorithm consistency in estimating feature importance will be crucial for future developments to ensure robustness and reliability in explainable AI frameworks. This work sets a solid foundation for ongoing advancements in making AI more interpretable and accountable to its users, paving the way for broader acceptance and integration of autonomous systems.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (44)
  1. Explainable artificial intelligence (xai): Concepts, taxonomies, opportunities and challenges toward responsible ai, Information fusion 58 (2020) 82–115.
  2. Trustworthy artificial intelligence: a review, ACM computing surveys (CSUR) 55 (2022) 1–38.
  3. I. Varošanec, On the path to the future: mapping the notion of transparency in the eu regulatory framework for ai, International Review of Law, Computers & Technology 36 (2022) 95–117.
  4. Explainability in ai policies: a critical review of communications, reports, regulations, and standards in the eu, us, and uk, in: Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, 2023, pp. 1198–1212.
  5. F. Doshi-Velez, B. Kim, Towards a rigorous science of interpretable machine learning, arXiv preprint arXiv:1702.08608 (2017).
  6. Interpreting black-box models: a review on explainable artificial intelligence, Cognitive Computation 16 (2024) 45–74.
  7. Explainable artificial intelligence (xai) post-hoc explainability methods: Risks and limitations in non-discrimination law, AI and Ethics 2 (2022) 815–826.
  8. Reliable post hoc explanations: Modeling uncertainty in explainability, Advances in neural information processing systems 34 (2021) 9391–9404.
  9. Post-hoc interpretability for neural nlp: A survey, ACM Computing Surveys 55 (2022) 1–42.
  10. Explainable ai: A review of machine learning interpretability methods, Entropy 23 (2020) 18.
  11. A survey on xai and natural language explanations, Information Processing & Management 60 (2023) 103111.
  12. Driving with llms: Fusing object-level vector modality for explainable autonomous driving, arXiv preprint arXiv:2310.01957 (2023).
  13. A survey on multimodal large language models for autonomous driving, in: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2024, pp. 958–979.
  14. M. D. Zeiler, R. Fergus, Visualizing and understanding convolutional networks, in: Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part I 13, Springer, 2014, pp. 818–833.
  15. Grad-cam: Visual explanations from deep networks via gradient-based localization, in: Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 618–626.
  16. Neural machine translation by jointly learning to align and translate, arXiv preprint arXiv:1409.0473 (2014).
  17. A survey on large language models: Applications, challenges, limitations, and practical usage, Authorea Preprints (2023).
  18. Are large language models post hoc explainers?, arXiv preprint arXiv:2310.05797 (2023).
  19. Fine-tuning large language model based explainable recommendation with explainable quality reward, in: Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, 2024, pp. 9250–9259.
  20. T. Miller, Explanation in artificial intelligence: Insights from the social sciences, Artificial intelligence 267 (2019) 1–38.
  21. D. J. Hilton, Conversational processes and causal explanation., Psychological Bulletin 107 (1990) 65.
  22. A. Cawsey, Planning interactive explanations, International Journal of Man-Machine Studies 38 (1993) 169–199.
  23. H. P. Grice, Logic and conversation, in: Speech acts, Brill, 1975, pp. 41–58.
  24. A taxonomy of social cues for conversational agents, International Journal of Human-Computer Studies 132 (2019) 138–161.
  25. I. A. Apperly, S. A. Butterfill, Do humans have two systems to track beliefs and belief-like states?, Psychological review 116 (2009) 953.
  26. R. R. Hoffman, G. Klein, Explaining explanation, part 1: theoretical foundations, IEEE Intelligent Systems 32 (2017) 68–73.
  27. Machine learning interpretability: A survey on methods and metrics, Electronics 8 (2019) 832.
  28. D. Hilton, Social attribution and explanation, Oxford Academic (2017).
  29. P. Lipton, Contrastive explanation, Royal Institute of Philosophy Supplements 27 (1990) 247–266.
  30. K. Epstude, N. J. Roese, The functional theory of counterfactual thinking, Personality and social psychology review 12 (2008) 168–192.
  31. Context-based image explanations for deep neural networks, Image and Vision Computing 116 (2021) 104310.
  32. Perturbation-based methods for explaining deep neural networks: A survey, Pattern Recognition Letters 150 (2021) 228–234.
  33. M. Robnik-Šikonja, M. Bohanec, Perturbation-based explanations of prediction models, Human and Machine Learning: Visible, Explainable, Trustworthy and Transparent (2018) 159–175.
  34. Explaining by removing: A unified framework for model explanation, Journal of Machine Learning Research 22 (2021) 1–90.
  35. From spoken thoughts to automated driving commentary: Predicting and explaining intelligent vehicles’ actions, in: 2022 IEEE Intelligent Vehicles Symposium (IV), IEEE, 2022, pp. 1040–1047.
  36. Visual explanations for dnns with contextual importance, in: Explainable and Transparent AI and Multi-Agent Systems: Third International Workshop, EXTRAAMAS 2021, Virtual Event, May 3–7, 2021, Revised Selected Papers 3, Springer, 2021, pp. 83–96.
  37. Retrieval-augmented generation for large language models: A survey, arXiv preprint arXiv:2312.10997 (2023).
  38. Real-time object detection and semantic segmentation for autonomous driving, in: MIPPR 2017: Automatic Target Recognition and Navigation, volume 10608, SPIE, 2018, pp. 167–174.
  39. Modeling and assessing an intelligent system for safety in human-robot collaboration using deep and machine learning techniques, Multimedia Tools and Applications (2021) 1–27.
  40. Hierarchical vertex regression-based segmentation of head and neck ct images for radiotherapy planning, IEEE Transactions on Image Processing 27 (2017) 923–937.
  41. Places: A 10 million image database for scene recognition, IEEE Transactions on Pattern Analysis and Machine Intelligence (2017).
  42. Encoder-decoder with atrous separable convolution for semantic image segmentation, in: Proceedings of the European conference on computer vision (ECCV), 2018, pp. 801–818.
  43. A brief survey on semantic segmentation with deep learning, Neurocomputing 406 (2020) 302–321.
  44. A survey on semantic segmentation, in: 2018 IEEE International Conference on Data Mining Workshops (ICDMW), IEEE, 2018, pp. 1233–1240.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Sule Tekkesinoglu (2 papers)
  2. Lars Kunze (40 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

Youtube Logo Streamline Icon: https://streamlinehq.com