Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

XToM: Exploring the Multilingual Theory of Mind for Large Language Models (2506.02461v1)

Published 3 Jun 2025 in cs.CL

Abstract: Theory of Mind (ToM), the ability to infer mental states in others, is pivotal for human social cognition. Existing evaluations of ToM in LLMs are largely limited to English, neglecting the linguistic diversity that shapes human cognition. This limitation raises a critical question: can LLMs exhibit Multilingual Theory of Mind, which is the capacity to reason about mental states across diverse linguistic contexts? To address this gap, we present XToM, a rigorously validated multilingual benchmark that evaluates ToM across five languages and incorporates diverse, contextually rich task scenarios. Using XToM, we systematically evaluate LLMs (e.g., DeepSeek R1), revealing a pronounced dissonance: while models excel in multilingual language understanding, their ToM performance varies across languages. Our findings expose limitations in LLMs' ability to replicate human-like mentalizing across linguistic contexts.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (17)
  1. Chunkit Chan (19 papers)
  2. Yauwai Yim (8 papers)
  3. Hongchuan Zeng (3 papers)
  4. Zhiying Zou (1 paper)
  5. Xinyuan Cheng (4 papers)
  6. Zhifan Sun (3 papers)
  7. Zheye Deng (12 papers)
  8. Kawai Chung (1 paper)
  9. Yuzhuo Ao (1 paper)
  10. Yixiang Fan (1 paper)
  11. Cheng Jiayang (11 papers)
  12. Ercong Nie (25 papers)
  13. Ginny Y. Wong (13 papers)
  14. Helmut Schmid (20 papers)
  15. Hinrich Schütze (250 papers)
  16. Simon See (74 papers)
  17. Yangqiu Song (196 papers)
Youtube Logo Streamline Icon: https://streamlinehq.com