XToM: Exploring the Multilingual Theory of Mind for Large Language Models (2506.02461v1)
Abstract: Theory of Mind (ToM), the ability to infer mental states in others, is pivotal for human social cognition. Existing evaluations of ToM in LLMs are largely limited to English, neglecting the linguistic diversity that shapes human cognition. This limitation raises a critical question: can LLMs exhibit Multilingual Theory of Mind, which is the capacity to reason about mental states across diverse linguistic contexts? To address this gap, we present XToM, a rigorously validated multilingual benchmark that evaluates ToM across five languages and incorporates diverse, contextually rich task scenarios. Using XToM, we systematically evaluate LLMs (e.g., DeepSeek R1), revealing a pronounced dissonance: while models excel in multilingual language understanding, their ToM performance varies across languages. Our findings expose limitations in LLMs' ability to replicate human-like mentalizing across linguistic contexts.
- Chunkit Chan (19 papers)
- Yauwai Yim (8 papers)
- Hongchuan Zeng (3 papers)
- Zhiying Zou (1 paper)
- Xinyuan Cheng (4 papers)
- Zhifan Sun (3 papers)
- Zheye Deng (12 papers)
- Kawai Chung (1 paper)
- Yuzhuo Ao (1 paper)
- Yixiang Fan (1 paper)
- Cheng Jiayang (11 papers)
- Ercong Nie (25 papers)
- Ginny Y. Wong (13 papers)
- Helmut Schmid (20 papers)
- Hinrich Schütze (250 papers)
- Simon See (74 papers)
- Yangqiu Song (196 papers)