Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Perceptions to Beliefs: Exploring Precursory Inferences for Theory of Mind in Large Language Models (2407.06004v3)

Published 8 Jul 2024 in cs.CL

Abstract: While humans naturally develop theory of mind (ToM), the capability to understand other people's mental states and beliefs, state-of-the-art LLMs underperform on simple ToM benchmarks. We posit that we can extend our understanding of LLMs' ToM abilities by evaluating key human ToM precursors$-$perception inference and perception-to-belief inference$-$in LLMs. We introduce two datasets, Percept-ToMi and Percept-FANToM, to evaluate these precursory inferences for ToM in LLMs by annotating characters' perceptions on ToMi and FANToM, respectively. Our evaluation of eight state-of-the-art LLMs reveals that the models generally perform well in perception inference while exhibiting limited capability in perception-to-belief inference (e.g., lack of inhibitory control). Based on these results, we present PercepToM, a novel ToM method leveraging LLMs' strong perception inference capability while supplementing their limited perception-to-belief inference. Experimental results demonstrate that PercepToM significantly enhances LLM's performance, especially in false belief scenarios.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Chani Jung (2 papers)
  2. Dongkwan Kim (25 papers)
  3. Jiho Jin (15 papers)
  4. Jiseon Kim (12 papers)
  5. Yeon Seonwoo (7 papers)
  6. Yejin Choi (287 papers)
  7. Alice Oh (81 papers)
  8. Hyunwoo Kim (52 papers)
Citations (2)