Reliable explainability for current AI models
Develop robust, reliable methods to properly explain the decisions of current artificial intelligence systems—especially deep learning and large language models—beyond existing post-hoc interpretability techniques whose outputs can be inconsistent or misleading.
Sponsor
References
In practice, this research confirms that we do not know how to properly explain the decisions of current AIs .
— The Impact of Artificial Intelligence on Human Thought
(2508.16628 - Gesnot, 15 Aug 2025) in Chapter 6: "Black Box" AI and the Hypothesis of an Orchestrating Consciousness, Explainability and Trust in AI