Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Theoretical Framework for AI Models Explainability with Application in Biomedicine (2212.14447v4)

Published 29 Dec 2022 in cs.AI, cs.CV, and cs.LG

Abstract: EXplainable Artificial Intelligence (XAI) is a vibrant research topic in the artificial intelligence community, with growing interest across methods and domains. Much has been written about the subject, yet XAI still lacks shared terminology and a framework capable of providing structural soundness to explanations. In our work, we address these issues by proposing a novel definition of explanation that is a synthesis of what can be found in the literature. We recognize that explanations are not atomic but the combination of evidence stemming from the model and its input-output mapping, and the human interpretation of this evidence. Furthermore, we fit explanations into the properties of faithfulness (i.e., the explanation being a true description of the model's inner workings and decision-making process) and plausibility (i.e., how much the explanation looks convincing to the user). Using our proposed theoretical framework simplifies how these properties are operationalized and it provides new insight into common explanation methods that we analyze as case studies.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Matteo Rizzo (5 papers)
  2. Alberto Veneri (2 papers)
  3. Andrea Albarelli (4 papers)
  4. Claudio Lucchese (22 papers)
  5. Marco Nobile (3 papers)
  6. Cristina Conati (15 papers)
Citations (4)