Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ToM-LM: Delegating Theory of Mind Reasoning to External Symbolic Executors in Large Language Models (2404.15515v3)

Published 23 Apr 2024 in cs.CL and cs.AI

Abstract: Theory of Mind (ToM) refers to the ability of individuals to attribute mental states to others. While LLMs have shown some promise with ToM ability, they still struggle with complex ToM reasoning. Our approach leverages an external symbolic executor, specifically the SMCDEL model checker, and fine-tuning to improve the ToM reasoning ability of LLMs. In our approach, an LLM is first fine-tuned through pairs of natural language and symbolic formulation representation of ToM problems and is then instructed to generate the symbolic formulation with a one-shot in-context example. The generated symbolic formulation is then executed by the SMCDEL model checker to perform transparent and verifiable ToM reasoning and give the final result. We demonstrate that our approach, ToM-LM, shows a significant improvement over all the constructed baselines. Our study proposes a novel view about externalizing a particular component of ToM reasoning, mainly reasoning about beliefs, and suggests generalizing it to other aspects of ToM reasoning.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (23)
  1. Distributed computing meets game theory: combining insights from two fields. ACM SIGACT News, 42(2):69–76, June 2011.
  2. Does the autistic child have a “theory of mind”? Cognition, 21(1):37–46, October 1985.
  3. Sparks of artificial general intelligence: Early experiments with gpt-4. arXiv preprint arXiv:2303.12712, 2023.
  4. Reasoning about knowledge. MIT press, 2004.
  5. The neural basis of mentalizing. Neuron, 50(4):531–534, May 2006.
  6. Joseph Y. Halpern. Substantive rationality and backward induction. Games and Economic Behavior, 37(2):425–435, November 2001.
  7. An epistemic characterization of zero knowledge. In Proceedings of the 12th Conference on Theoretical Aspects of Rationality and Knowledge, TARK ’09, page 156–165, New York, NY, USA, July 2009. Association for Computing Machinery.
  8. Unveiling theory of mind in large language models: A parallel to single neurons in the human brain.
  9. Large language models are zero-shot reasoners. (arXiv:2205.11916), January 2023. arXiv:2205.11916 [cs].
  10. Michal Kosinski. Evaluating large language models in theory of mind tasks. 2023.
  11. Theory of mind for multi-agent collaboration via large language models. (arXiv:2310.10701), October 2023. arXiv:2310.10701 [cs].
  12. Towards a holistic landscape of situated theory of mind in large language models. (arXiv:2310.19619), October 2023. arXiv:2310.19619 [cs].
  13. Boosting theory-of-mind performance in large language models via prompting.
  14. Linc: A neurosymbolic approach for logical reasoning by combining language models with first-order logic provers. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, page 5153–5176, Singapore, 2023. Association for Computational Linguistics.
  15. Logic-lm: Empowering large language models with symbolic solvers for faithful logical reasoning. (arXiv:2305.12295), October 2023. arXiv:2305.12295 [cs].
  16. Does the chimpanzee have a theory of mind? Behavioral and Brain Sciences, 1(4):515–526, December 1978.
  17. What do theory-of-mind tasks actually measure? theory and practice. Perspectives on Psychological Science, 15(2):384–396, March 2020.
  18. Neural theory-of-mind? on the limits of social intelligence in large lms. (arXiv:2210.13312), April 2023. arXiv:2210.13312 [cs].
  19. Toolformer: Language models can teach themselves to use tools. (arXiv:2302.04761), February 2023. arXiv:2302.04761 [cs].
  20. Mindgames: Targeting theory of mind in large language models with dynamic epistemic modal logic. (arXiv:2305.03353), November 2023. arXiv:2305.03353 [cs].
  21. Do large language models know what humans know? Cognitive Science, 47(7):e13309, 2023.
  22. Symbolic model checking for dynamic epistemic logic — s5 and beyond*. Journal of Logic and Computation, 28(2):367–402, March 2018.
  23. Chain-of-thought prompting elicits reasoning in large language models. 2022.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Weizhi Tang (4 papers)
  2. Vaishak Belle (59 papers)
Citations (1)