Papers
Topics
Authors
Recent
Search
2000 character limit reached

Language models can learn implicit multi-hop reasoning, but only if they have lots of training data

Published 23 May 2025 in cs.CL | (2505.17923v1)

Abstract: Implicit reasoning is the ability of a LLM to solve multi-hop reasoning tasks in a single forward pass, without chain of thought. We investigate this capability using GPT2-style LLMs trained from scratch on controlled $k$-hop reasoning datasets ($k = 2, 3, 4$). We show that while such models can indeed learn implicit $k$-hop reasoning, the required training data grows exponentially in $k$, and the required number of transformer layers grows linearly in $k$. We offer a theoretical explanation for why this depth growth is necessary. We further find that the data requirement can be mitigated, but not eliminated, through curriculum learning.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 4 likes about this paper.