2000 character limit reached
First Heuristic Then Rational: Dynamic Use of Heuristics in Language Model Reasoning (2406.16078v2)
Published 23 Jun 2024 in cs.CL
Abstract: Multi-step reasoning instruction, such as chain-of-thought prompting, is widely adopted to explore better LLMs (LMs) performance. We report on the systematic strategy that LMs employ in such a multi-step reasoning process. Our controlled experiments reveal that LMs rely more heavily on heuristics, such as lexical overlap, in the earlier stages of reasoning, where more reasoning steps remain to reach a goal. Conversely, their reliance on heuristics decreases as LMs progress closer to the final answer through multiple reasoning steps. This suggests that LMs can backtrack only a limited number of future steps and dynamically combine heuristic strategies with rationale ones in tasks involving multi-step reasoning.
- Yoichi Aoki (5 papers)
- Keito Kudo (7 papers)
- Tatsuki Kuribayashi (31 papers)
- Shusaku Sone (5 papers)
- Masaya Taniguchi (4 papers)
- Keisuke Sakaguchi (44 papers)
- Kentaro Inui (119 papers)