Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

General Intelligence Requires Reward-based Pretraining (2502.19402v2)

Published 26 Feb 2025 in cs.LG

Abstract: LLMs have demonstrated impressive real-world utility, exemplifying artificial useful intelligence (AUI). However, their ability to reason adaptively and robustly -- the haLLMarks of artificial general intelligence (AGI) -- remains fragile. While LLMs seemingly succeed in commonsense reasoning, programming, and mathematics, they struggle to generalize algorithmic understanding across novel contexts. Our experiments with algorithmic tasks in esoteric programming languages reveal that LLM's reasoning overfits to the training data and is limited in its transferability. We hypothesize that the core issue underlying such limited transferability is the coupling of reasoning and knowledge in LLMs. To transition from AUI to AGI, we propose disentangling knowledge and reasoning through three key directions: (1) pretaining to reason using RL from scratch as an alternative to the widely used next-token prediction pretraining, (2) using a curriculum of synthetic tasks to ease the learning of a reasoning prior for RL that can then be transferred to natural language tasks, and (3) learning more generalizable reasoning functions using a small context window to reduce exploiting spurious correlations between tokens. Such a reasoning system coupled with a trained retrieval system and a large external memory bank as a knowledge store can overcome several limitations of existing architectures at learning to reason in novel scenarios.

Disentangling Knowledge and Reasoning in LLMs

The paper "General Reasoning Requires Learning to Reason from the Get-go" addresses the limitations of current LLMs in achieving true artificial general intelligence (AGI). The authors argue that a key limitation of existing LLMs is their reliance on next-token prediction during pretraining, which is predominantly based on passively collected internet data. This method inhibits their generalization potential across novel scenarios due to their intertwined nature of knowledge and reasoning. The paper presents compelling evidence and arguments for disentangling reasoning from knowledge in order to facilitate more robust and adaptable reasoning capabilities.

Key Contributions

  1. Evaluation of Current LLMs' Reasoning Capabilities: Through extensive evaluation using algorithmic tasks set in esoteric programming languages, the authors highlight deficiencies in the reasoning and generalization abilities of current state-of-the-art models. The tasks were designed to separate reasoning capabilities from rote memorization and were presented in unfamiliar syntaxes to remove bias from prior knowledge. The results showed that existing models, including advanced ones like o1, struggle with these tasks, highlighting their limitations in transferring learned reasoning skills to new contexts.
  2. Proposal for Reinforcement Learning (RL) Pretraining: The authors suggest that current paradigms involving supervised pretraining before RL finetuning hinder the model's ability to learn more generalized reasoning strategies, akin to the difference between AlphaGo and AlphaZero in their approach to Go. They propose integrating RL from the onset of pretraining to avoid getting stuck in local minima of reasoning capabilities, drawing on how AlphaZero achieved superior performance by eschewing human demonstration-based pretraining entirely.
  3. Use of Synthetic Tasks for Reasoning Pretraining: Due to the complex exploration space of natural language, the paper advocates for the use of synthetic tasks with reduced token spaces to pretrain reasoning abilities. By focusing on tasks that encapsulate universal reasoning primitives, the authors hypothesize that this approach could scaffold learning more complex reasoning required in natural language tasks.
  4. Architectural Enhancements to Disentangle Memory and Reasoning: The authors propose architectural changes to separate knowledge storage from reasoning processes. They suggest a model with an external semantic memory bank and a reasoning network working over a limited context window. This separation aims to reduce reliance on spurious correlations and to foster a more generalizable reasoning process. They draw on insights from cognitive science, particularly the concept of chunking that aids in handling complex information with limited working memory.

Implications and Future Directions

The proposals in this paper hold significant implications for the future of AI research, particularly in the pursuit of artificial general intelligence. By suggesting methods to better incorporate reasoning from the start of a model's training, the authors open avenues for overcoming current limitations observed in LLMs. Their approach of using synthetic tasks for developing reasoning prior within a limited token space provides a new direction for preparing models to handle larger, more complex natural language contexts effectively.

Furthermore, the architectural suggestion to decouple knowledge from reasoning supports a modular approach to model design, where reasoning mechanisms could potentially be standardized and reused across varied domains with different knowledge bases. Such a separation could facilitate easier updates and integration of new information without extensive retraining, mirroring human-like adaptability in AI systems.

The arguments and findings presented warrant further exploration into each of the proposed paradigm shifts. Future work would benefit from empirical validation of these hypotheses, particularly in larger, more realistic scenarios and varied domains, to truly assess the impact and scalability of disentangling reasoning and knowledge in the march towards more generalized AI systems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Seungwook Han (11 papers)
  2. Jyothish Pari (10 papers)
  3. Samuel J. Gershman (25 papers)
  4. Pulkit Agrawal (103 papers)