Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Token Transformation Matters: Towards Faithful Post-hoc Explanation for Vision Transformer (2403.14552v1)

Published 21 Mar 2024 in cs.CV

Abstract: While Transformers have rapidly gained popularity in various computer vision applications, post-hoc explanations of their internal mechanisms remain largely unexplored. Vision Transformers extract visual information by representing image regions as transformed tokens and integrating them via attention weights. However, existing post-hoc explanation methods merely consider these attention weights, neglecting crucial information from the transformed tokens, which fails to accurately illustrate the rationales behind the models' predictions. To incorporate the influence of token transformation into interpretation, we propose TokenTM, a novel post-hoc explanation method that utilizes our introduced measurement of token transformation effects. Specifically, we quantify token transformation effects by measuring changes in token lengths and correlations in their directions pre- and post-transformation. Moreover, we develop initialization and aggregation rules to integrate both attention weights and token transformation effects across all layers, capturing holistic token contributions throughout the model. Experimental results on segmentation and perturbation tests demonstrate the superiority of our proposed TokenTM compared to state-of-the-art Vision Transformer explanation methods.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. Quantifying attention flow in transformers. In ACL, 2020.
  2. Openxai: Towards a transparent evaluation of model explanations. In NeurIPS, 2022.
  3. Xai for transformers: Better explanations through conservative propagation. In ICML, 2022.
  4. A diagnostic study of explainability techniques for text classification. In EMNLP, 2020.
  5. On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. In PloS one, 10(7):e0130140, 2015.
  6. Grad-sam: Explaining transformers via gradient self-attention maps. In CIKM, 2021.
  7. Visual explanations via iterated integrated attributions. In ICCV, 2023a.
  8. Deep integrated explanations. In CIKM, 2023b.
  9. Layer-wise relevance propagation for neural networks with local renormalization layers. In ICANN, 2016.
  10. On identifiability in transformers. In ICLR, 2020.
  11. End-to-end object detection with transformers. In ECCV, 2020.
  12. Generic attention-model explainability for interpreting bi-modal and encoder-decoder transformers. In ICCV, 2021a.
  13. Transformer interpretability beyond attention visualization. In CVPR, 2021b.
  14. Generating hierarchical explanations on text classification via feature interaction detection. In ACL, 2020.
  15. Real time image saliency for black box classifiers. In NeurIPS, 2017.
  16. Eraser: A benchmark to evaluate rationalized nlp models. In ACL, 2020.
  17. An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR, 2021.
  18. Kawin Ethayarajh. How contextual are contextualized word representations? comparing the geometry of bert, elmo, and gpt-2 embeddings. In EMNLP-IJCNLP, 2019.
  19. Interpretable explanations of black boxes by meaningful perturbation. In ICCV, 2017.
  20. Understanding individual decisions of cnns via contrastive backpropagation. In ACCV, 2018.
  21. Imagenet auto-annotation with segmentation propagation. IJCV, 110:328–348, 2014.
  22. Visualization of supervised and self-supervised neural networks via attribution guided factorization. In AAAI, 2021.
  23. Deep residual learning for image recognition. In CVPR, 2016.
  24. Masked autoencoders are scalable vision learners. In CVPR, 2022.
  25. Explaining convolutional neural networks using softmax gradient layer-wise relevance propagation. In ICCV Workshop, 2019.
  26. Attention is not explanation. In NAACL, 2019.
  27. Attention is not only a weight: Analyzing transformers with vector norms. In EMNLP, 2020.
  28. Learning multiple layers of features from tiny images. 2009.
  29. Adaptive cross-modal prototypes for cross-domain visual-language retrieval. In CVPR, 2021.
  30. Swin transformer v2: Scaling up capacity and resolution. In CVPR, 2022.
  31. A unified approach to interpreting model predictions. In NeurIPS, 2017.
  32. Visualizing deep convolutional neural networks using natural pre-images. IJCV, 120:233–255, 2016.
  33. Explaining nonlinear classification decisions with deep taylor decomposition. PR, 65:211–222, 2017.
  34. Dong Nguyen. Comparing automatic and human evaluation of local explanations for text classification. In NAACL, 2018.
  35. Attcat: Explaining transformers via attentive class activation tokens. In NeurIPS, 2022.
  36. Imagenet large scale visual recognition challenge. IJCV, 115:211–252, 2015.
  37. Grad-cam: Visual explanations from deep networks via gradient-based localization. In ICCV, 2017.
  38. Is attention interpretable? In ACL, 2019.
  39. Learning important features through propagating activation differences. In ICML, 2017a.
  40. Not just a black box: Learning important features through propagating activation differences. In ICML, 2017b.
  41. Smoothgrad: removing noise by adding noise. In ICML Workshop, 2017.
  42. Striving for simplicity: The all convolutional net. In ICLRW, 2015.
  43. Full-gradient representation for neural network visualization. In NeurIPS, 2019.
  44. Gilbert Strang. Linear algebra and its applications. Belmont, CA: Thomson, Brooks/Cole, 2006.
  45. Axiomatic attribution for deep networks. In ICML, 2017.
  46. Training data-efficient image transformers & distillation through attention. In ICML, 2021.
  47. Attention is all you need. In NeurIPS, 2017.
  48. Analyzing multi-head self-attention: Specialized heads do the heavy lifting, the rest can be pruned. In ACL, 2019.
  49. Attention is not not explanation. In EMNLP, 2019.
  50. Visualizing and understanding convolutional networks. In ECCV, 2014.
  51. Learning deep features for discriminative localization. In CVPR, 2016.
  52. Interpreting deep visual representations via network dissection. IEEE TPAMI, 41(9):2131–2145, 2018.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Junyi Wu (15 papers)
  2. Bin Duan (22 papers)
  3. Weitai Kang (11 papers)
  4. Hao Tang (379 papers)
  5. Yan Yan (242 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.

Youtube Logo Streamline Icon: https://streamlinehq.com