Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Interpreting Deep Learning Models in Natural Language Processing: A Review (2110.10470v2)

Published 20 Oct 2021 in cs.CL

Abstract: Neural network models have achieved state-of-the-art performances in a wide range of NLP tasks. However, a long-standing criticism against neural network models is the lack of interpretability, which not only reduces the reliability of neural NLP systems but also limits the scope of their applications in areas where interpretability is essential (e.g., health care applications). In response, the increasing interest in interpreting neural NLP models has spurred a diverse array of interpretation methods over recent years. In this survey, we provide a comprehensive review of various interpretation methods for neural models in NLP. We first stretch out a high-level taxonomy for interpretation methods in NLP, i.e., training-based approaches, test-based approaches, and hybrid approaches. Next, we describe sub-categories in each category in detail, e.g., influence-function based methods, KNN-based methods, attention-based models, saliency-based methods, perturbation-based methods, etc. We point out deficiencies of current methods and suggest some avenues for future research.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Xiaofei Sun (36 papers)
  2. Diyi Yang (151 papers)
  3. Xiaoya Li (42 papers)
  4. Tianwei Zhang (199 papers)
  5. Yuxian Meng (37 papers)
  6. Han Qiu (60 papers)
  7. Guoyin Wang (108 papers)
  8. Eduard Hovy (115 papers)
  9. Jiwei Li (137 papers)
Citations (42)

Summary

We haven't generated a summary for this paper yet.