Reliable explainability for current AI models
Develop robust, reliable methods to properly explain the decisions of current artificial intelligence systems—especially deep learning and large language models—beyond existing post-hoc interpretability techniques whose outputs can be inconsistent or misleading.
References
In practice, this research confirms that we do not know how to properly explain the decisions of current AIs .
— The Impact of Artificial Intelligence on Human Thought
(2508.16628 - Gesnot, 15 Aug 2025) in Chapter 6: "Black Box" AI and the Hypothesis of an Orchestrating Consciousness, Explainability and Trust in AI