Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

All Languages Matter: On the Multilingual Safety of Large Language Models (2310.00905v2)

Published 2 Oct 2023 in cs.CL and cs.AI

Abstract: Safety lies at the core of developing and deploying LLMs. However, previous safety benchmarks only concern the safety in one language, e.g. the majority language in the pretraining data such as English. In this work, we build the first multilingual safety benchmark for LLMs, XSafety, in response to the global deployment of LLMs in practice. XSafety covers 14 kinds of commonly used safety issues across 10 languages that span several language families. We utilize XSafety to empirically study the multilingual safety for 4 widely-used LLMs, including both close-API and open-source models. Experimental results show that all LLMs produce significantly more unsafe responses for non-English queries than English ones, indicating the necessity of developing safety alignment for non-English languages. In addition, we propose several simple and effective prompting methods to improve the multilingual safety of ChatGPT by evoking safety knowledge and improving cross-lingual generalization of safety alignment. Our prompting method can significantly reduce the ratio of unsafe responses from 19.1% to 9.7% for non-English queries. We release our data at https://github.com/Jarviswang94/Multilingual_safety_benchmark.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Wenxuan Wang (128 papers)
  2. Zhaopeng Tu (135 papers)
  3. Chang Chen (48 papers)
  4. Youliang Yuan (18 papers)
  5. Jen-tse Huang (46 papers)
  6. Wenxiang Jiao (44 papers)
  7. Michael R. Lyu (176 papers)
Citations (27)
Github Logo Streamline Icon: https://streamlinehq.com