Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

An Experimental Investigation into the Evaluation of Explainability Methods (2305.16361v1)

Published 25 May 2023 in cs.LG, cs.AI, and cs.CV

Abstract: EXplainable Artificial Intelligence (XAI) aims to help users to grasp the reasoning behind the predictions of an AI system. Many XAI approaches have emerged in recent years. Consequently, a subfield related to the evaluation of XAI methods has gained considerable attention, with the aim to determine which methods provide the best explanation using various approaches and criteria. However, the literature lacks a comparison of the evaluation metrics themselves, that one can use to evaluate XAI methods. This work aims to fill this gap by comparing 14 different metrics when applied to nine state-of-the-art XAI methods and three dummy methods (e.g., random saliency maps) used as references. Experimental results show which of these metrics produces highly correlated results, indicating potential redundancy. We also demonstrate the significant impact of varying the baseline hyperparameter on the evaluation metric values. Finally, we use dummy methods to assess the reliability of metrics in terms of ranking, pointing out their limitations.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Sédrick Stassin (1 paper)
  2. Alexandre Englebert (3 papers)
  3. Julien Albert (2 papers)
  4. Nassim Versbraegen (2 papers)
  5. Gilles Peiffer (1 paper)
  6. Miriam Doh (7 papers)
  7. Nicolas Riche (5 papers)
  8. Christophe De Vleeschouwer (52 papers)
  9. Géraldin Nanfack (1 paper)
  10. Benoît Frenay (1 paper)
Citations (5)

Summary

We haven't generated a summary for this paper yet.