Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Order in the Court: Explainable AI Methods Prone to Disagreement (2105.03287v3)

Published 7 May 2021 in cs.LG and cs.CL

Abstract: By computing the rank correlation between attention weights and feature-additive explanation methods, previous analyses either invalidate or support the role of attention-based explanations as a faithful and plausible measure of salience. To investigate whether this approach is appropriate, we compare LIME, Integrated Gradients, DeepLIFT, Grad-SHAP, Deep-SHAP, and attention-based explanations, applied to two neural architectures trained on single- and pair-sequence language tasks. In most cases, we find that none of our chosen methods agree. Based on our empirical observations and theoretical objections, we conclude that rank correlation does not measure the quality of feature-additive methods. Practitioners should instead use the numerous and rigorous diagnostic methods proposed by the community.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Michael Neely (5 papers)
  2. Stefan F. Schouten (5 papers)
  3. Maurits J. R. Bleeker (1 paper)
  4. Ana Lucic (15 papers)
Citations (38)

Summary

We haven't generated a summary for this paper yet.