Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
120 tokens/sec
GPT-4o
10 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
3 tokens/sec
DeepSeek R1 via Azure Pro
55 tokens/sec
2000 character limit reached

Can Stories Help LLMs Reason? Curating Information Space Through Narrative (2410.19221v1)

Published 25 Oct 2024 in cs.CL, cs.AI, and cs.LG

Abstract: Narratives are widely recognized as a powerful tool for structuring information and facilitating comprehension of complex ideas in various domains such as science communication. This paper investigates whether incorporating narrative elements can assist LLMs in solving complex problems more effectively. We propose a novel approach, Story of Thought (SoT), integrating narrative structures into prompting techniques for problem-solving. This approach involves constructing narratives around problem statements and creating a framework to identify and organize relevant information. Our experiments show that using various LLMs with SoT consistently surpasses using them with other techniques on physics, chemistry, math, and biology questions in both the GPQA and JEEBench datasets. The narrative-based information curation process in SoT enhances problem comprehension by contextualizing critical in-domain information and highlighting causal relationships within the problem space.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (42)
  1. H Porter Abbott. 2020. The Cambridge introduction to narrative. Cambridge University Press.
  2. Have LLMs advanced enough? a challenging problem solving benchmark for large language models. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 7527–7543, Singapore. Association for Computational Linguistics.
  3. Graph of Thoughts: Solving Elaborate Problems with Large Language Models. Proceedings of the AAAI Conference on Artificial Intelligence, 38(16):17682–17690.
  4. Gordon H Bower and Michal C Clark. 1969. Narrative stories as mediators for serial learning. Psychonomic science, 14(4):181–182.
  5. Jerome Bruner. 1991. The narrative construction of reality. Critical inquiry, 18(1):1–21.
  6. The design narrative in design learning: Adjusting the inertia of attention and enhancing design integrity. The Design Journal, 26(4):519–535.
  7. Training verifiers to solve math word problems. arXiv preprint arXiv:2110.14168.
  8. Michael F Dahlstrom. 2014. Using narratives and storytelling to communicate science with nonexpert audiences. Proceedings of the national academy of sciences, 111(supplement_4):13614–13620.
  9. The use of narrative in science and health communication: a scoping review. Patient Education and Counseling, page 107752.
  10. Rethinking narrative: Leveraging storytelling for science learning. Childhood Education, 94(6):4–12.
  11. Walter R Fisher. 2021. Human communication as narration: Toward a philosophy of reason, value, and action. University of South Carolina Press.
  12. Jonathan Gottschall. 2012. The storytelling animal: How stories make us human. Houghton Mifflin Harcourt.
  13. Looking beyond the surface: A challenge set for reading comprehension over multiple sentences. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 252–262, New Orleans, Louisiana. Association for Computational Linguistics.
  14. The NarrativeQA reading comprehension challenge. Transactions of the Association for Computational Linguistics, 6:317–328.
  15. TellMeWhy: A dataset for answering why-questions in narratives. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 596–610, Online. Association for Computational Linguistics.
  16. Making language models better reasoners with step-aware verifier. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 5315–5333, Toronto, Canada. Association for Computational Linguistics.
  17. Program induction by rationale generation: Learning to solve and explain algebraic word problems. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 158–167, Vancouver, Canada. Association for Computational Linguistics.
  18. Learn to explain: Multimodal reasoning via thought chains for science question answering. In The 36th Conference on Neural Information Processing Systems (NeurIPS).
  19. Susana Martinez-Conde and Stephen L Macknik. 2017. Finding the plot in science storytelling in hopes of enhancing science communication. Proceedings of the National Academy of Sciences, 114(31):8127–8129.
  20. Systematic literature review on narrative-based learning in educational technology learning environments (2007-2017).
  21. Can a suit of armor conduct electricity? a new dataset for open book question answering. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2381–2391, Brussels, Belgium. Association for Computational Linguistics.
  22. Rethinking the role of demonstrations: What makes in-context learning work? In EMNLP.
  23. A corpus and cloze evaluation for deeper understanding of commonsense stories. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 839–849, San Diego, California. Association for Computational Linguistics.
  24. Aquiles Negrete and Cecilia Lartigue. 2004. Learning from education to communicate science as a good story. Endeavour, 28(3):120–124.
  25. A theoretical framework for narrative explanation in science. Science education, 89(4):535–563.
  26. Medmcqa: A large-scale multi-subject multi-choice dataset for medical domain question answering. In Proceedings of the Conference on Health, Inference, and Learning, volume 174 of Proceedings of Machine Learning Research, pages 248–260. PMLR.
  27. Reasoning with language model prompting: A survey. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 5368–5393, Toronto, Canada. Association for Computational Linguistics.
  28. GPQA: A graduate-level google-proof q&a benchmark. In First Conference on Language Modeling.
  29. Unveiling information through narrative in conversational information seeking. In Proceedings of the 6th ACM Conference on Conversational User Interfaces, CUI ’24, New York, NY, USA. Association for Computing Machinery.
  30. Joanna Szurmak and Mindy Thuna. 2013. Tell me a story: The use of narrative as tool for instruction.
  31. Scibench: Evaluating college-level scientific problem-solving abilities of large language models. arXiv preprint arXiv:2307.10635.
  32. Self-consistency improves chain of thought reasoning in language models. ArXiv, abs/2203.11171.
  33. Chain-of-thought prompting elicits reasoning in large language models. Advances in neural information processing systems, 35:24824–24837.
  34. Crowdsourcing multiple choice science questions. In Proceedings of the 3rd Workshop on Noisy User-generated Text, pages 94–106, Copenhagen, Denmark. Association for Computational Linguistics.
  35. Constructing datasets for multi-hop reading comprehension across documents. Transactions of the Association for Computational Linguistics, 6:287–302.
  36. A broad-coverage challenge corpus for sentence understanding through inference. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 1112–1122, New Orleans, Louisiana. Association for Computational Linguistics.
  37. Daniel T Willingham. 2004. Ask the cognitive scientist the privileged status of story. American Educator, 28:43–45.
  38. Beyond chain-of-thought: A survey of chain-of-x paradigms for llms. arXiv preprint arXiv:2404.15676.
  39. Tree of Thoughts: Deliberate problem solving with large language models. Preprint, arXiv:2305.10601.
  40. Large language models as analogical reasoners. arXiv preprint arXiv:2310.01714.
  41. The unreliability of explanations in few-shot prompting for textual reasoning. In Proceedings of the 36th International Conference on Neural Information Processing Systems, NIPS ’22, Red Hook, NY, USA. Curran Associates Inc.
  42. Paul J Zak. 2015. Why inspiring stories make us react: The neuroscience of narrative. In Cerebrum: the Dana forum on brain science, volume 2015. Dana Foundation.

Summary

  • The paper introduces the Story of Thought approach that integrates narrative techniques into LLM reasoning, significantly boosting problem-solving accuracy.
  • It employs a three-step pipeline—question clarification, narrative generation, and problem solving—to outperform traditional zero-shot and CoT methods.
  • Empirical results on GPQA and JEEBench datasets demonstrate that narrative-driven prompts improve LLM comprehension and contextual reasoning.

Can Stories Help LLMs Reason? An Analysis of the Story of Thought Approach

The paper Can Stories Help LLMs Reason? Curating Information Space Through Narrative presents a novel methodology aimed at enhancing the problem-solving capabilities of LLMs by integrating narrative elements into their reasoning processes. Notably, this research introduces the "Story of Thought" (SoT) approach, which structures problem-solving prompts within narrative frameworks to improve comprehension and contextual understanding.

Research Context and Inquiry

The authors address the potential of narratives, widely acknowledged for their ability to elucidate and structure complex concepts in human cognition, as a means to enhance LLM reasoning. Building upon the established cognitive benefits of storytelling, this paper aims to determine if similar benefits can be conferred upon LLMs. The paper poses two key research questions: whether LLMs can generate coherent narratives around problem statements effectively, and if narrative incorporation can boost model performance in complex problem-solving scenarios.

Methodological Approach

The SoT method comprises a three-step pipeline designed to improve the reasoning and problem-solving proficiency of LLMs by fostering narrative thinking:

  1. Question Clarification: LLMs are tasked with breaking down and clarifying complex questions into sub-components, effectively serving as explorers to identify relevant subject areas for further narrative development.
  2. Narrative Generation: Using various narrative techniques, such as progressive disclosure, analogy, and metaphor, the LLMs generate structured narrative explanations of the clarified problem components. These narratives aim to provide context, relational reasoning, and cognitive insights, guiding the LLMs in understanding the problem space comprehensively.
  3. Problem Solving: The narrative-based understanding developed in the previous step is then employed to support the LLMs in addressing the task at hand, with an emphasis on integrating narrative comprehension into the reasoning process.

Experimental Evaluation

The paper evaluates the efficacy of the SoT approach on two challenging datasets: GPQA and JEEBench, both of which consist of complex, domain-specific questions requiring deep comprehension and reasoning. The paper benchmarks the SoT approach against various other prompting strategies, such as zero-shot and chain-of-thought (CoT) prompting.

The results reveal that the SoT method outperforms traditional prompting techniques across multiple LLMs, including models by Meta, Mistral, OpenAI, and Microsoft. Specifically, SoT displays significant accuracy improvements, notably enhancing the capabilities of models like Llama 3 70B and GPT-4 compared to their zero-shot and CoT counterparts. This suggests that narrative-driven prompts may effectively support LLMs in parsing and solving intricate problems by organizing information within a coherent context.

Implications and Future Directions

The findings emphasize the potential of narrative integration in LLM reasoning processes, offering a promising avenue for enhancing computational models' problem-solving aptitudes. The narrative-based prompting could revolutionize how LLMs are trained and leveraged in fields requiring complex decision-making and contextual understanding, such as science education and communication.

The paper opens new pathways for examining how narratives affect artificial reasoning and the broader practical implications for LLM adaptation in various applications. Further research could explore additional narrative techniques, cross-context efficacy, and their scalability across different domains and problem types. Additionally, the development of more refined methods for evaluating narrative quality and effectiveness could further enhance our understanding of narratives' role in AI reasoning.

In conclusion, the integration of narrative structures in LLM prompting as proposed in this paper shows promise as a technique for improving LLM reasoning capabilities, suggesting that storytelling's cognitive advantages can be systematically leveraged within AI frameworks to achieve superior problem-solving outcomes.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com