Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Interpreting the Robustness of Neural NLP Models to Textual Perturbations (2110.07159v2)

Published 14 Oct 2021 in cs.CL

Abstract: Modern NLP models are known to be sensitive to input perturbations and their performance can decrease when applied to real-world, noisy data. However, it is still unclear why models are less robust to some perturbations than others. In this work, we test the hypothesis that the extent to which a model is affected by an unseen textual perturbation (robustness) can be explained by the learnability of the perturbation (defined as how well the model learns to identify the perturbation with a small amount of evidence). We further give a causal justification for the learnability metric. We conduct extensive experiments with four prominent NLP models -- TextRNN, BERT, RoBERTa and XLNet -- over eight types of textual perturbations on three datasets. We show that a model which is better at identifying a perturbation (higher learnability) becomes worse at ignoring such a perturbation at test time (lower robustness), providing empirical support for our hypothesis.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Yunxiang Zhang (22 papers)
  2. Liangming Pan (59 papers)
  3. Samson Tan (21 papers)
  4. Min-Yen Kan (92 papers)
Citations (19)

Summary

We haven't generated a summary for this paper yet.