Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Assessing Adversarial Robustness of Large Language Models: An Empirical Study (2405.02764v2)

Published 4 May 2024 in cs.CL and cs.LG

Abstract: LLMs have revolutionized natural language processing, but their robustness against adversarial attacks remains a critical concern. We presents a novel white-box style attack approach that exposes vulnerabilities in leading open-source LLMs, including Llama, OPT, and T5. We assess the impact of model size, structure, and fine-tuning strategies on their resistance to adversarial perturbations. Our comprehensive evaluation across five diverse text classification tasks establishes a new benchmark for LLM robustness. The findings of this study have far-reaching implications for the reliable deployment of LLMs in real-world applications and contribute to the advancement of trustworthy AI systems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Zeyu Yang (27 papers)
  2. Zhao Meng (14 papers)
  3. Xiaochen Zheng (29 papers)
  4. Roger Wattenhofer (212 papers)
Citations (4)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets