Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Robustifying Language Models with Test-Time Adaptation (2310.19177v1)

Published 29 Oct 2023 in cs.CL and cs.LG

Abstract: Large-scale LLMs achieved state-of-the-art performance over a number of language tasks. However, they fail on adversarial language examples, which are sentences optimized to fool the LLMs but with similar semantic meanings for humans. While prior work focuses on making the LLM robust at training time, retraining for robustness is often unrealistic for large-scale foundation models. Instead, we propose to make the LLMs robust at test time. By dynamically adapting the input sentence with predictions from masked words, we show that we can reverse many language adversarial attacks. Since our approach does not require any training, it works for novel tasks at test time and can adapt to novel adversarial corruptions. Visualizations and empirical results on two popular sentence classification datasets demonstrate that our method can repair adversarial language attacks over 65% o

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Noah Thomas McDermott (1 paper)
  2. Junfeng Yang (80 papers)
  3. Chengzhi Mao (38 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.