Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

The Tower of Babel Revisited: Multilingual Jailbreak Prompts on Closed-Source Large Language Models (2505.12287v1)

Published 18 May 2025 in cs.CL and cs.AI

Abstract: LLMs have seen widespread applications across various domains, yet remain vulnerable to adversarial prompt injections. While most existing research on jailbreak attacks and hallucination phenomena has focused primarily on open-source models, we investigate the frontier of closed-source LLMs under multilingual attack scenarios. We present a first-of-its-kind integrated adversarial framework that leverages diverse attack techniques to systematically evaluate frontier proprietary solutions, including GPT-4o, DeepSeek-R1, Gemini-1.5-Pro, and Qwen-Max. Our evaluation spans six categories of security contents in both English and Chinese, generating 38,400 responses across 32 types of jailbreak attacks. Attack success rate (ASR) is utilized as the quantitative metric to assess performance from three dimensions: prompt design, model architecture, and language environment. Our findings suggest that Qwen-Max is the most vulnerable, while GPT-4o shows the strongest defense. Notably, prompts in Chinese consistently yield higher ASRs than their English counterparts, and our novel Two-Sides attack technique proves to be the most effective across all models. This work highlights a dire need for language-aware alignment and robust cross-lingual defenses in LLMs, and we hope it will inspire researchers, developers, and policymakers toward more robust and inclusive AI systems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Linghan Huang (5 papers)
  2. Haolin Jin (9 papers)
  3. Zhaoge Bi (1 paper)
  4. Pengyue Yang (1 paper)
  5. Peizhou Zhao (2 papers)
  6. Taozhao Chen (1 paper)
  7. Xiongfei Wu (4 papers)
  8. Lei Ma (195 papers)
  9. Huaming Chen (38 papers)