Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
112 tokens/sec
GPT-4o
11 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
4 tokens/sec
DeepSeek R1 via Azure Pro
33 tokens/sec
2000 character limit reached

Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs (2412.21187v2)

Published 30 Dec 2024 in cs.CL

Abstract: The remarkable performance of models like the OpenAI o1 can be attributed to their ability to emulate human-like long-time thinking during inference. These models employ extended chain-of-thought (CoT) processes, exploring multiple strategies to enhance problem-solving capabilities. However, a critical question remains: How to intelligently and efficiently scale computational resources during testing. This paper presents the first comprehensive study on the prevalent issue of overthinking in these models, where excessive computational resources are allocated for simple problems with minimal benefit. We introduce novel efficiency metrics from both outcome and process perspectives to evaluate the rational use of computational resources by o1-like models. Using a self-training paradigm, we propose strategies to mitigate overthinking, streamlining reasoning processes without compromising accuracy. Experimental results show that our approach successfully reduces computational overhead while preserving model performance across a range of testsets with varying difficulty levels, such as GSM8K, MATH500, GPQA, and AIME.

Summary

  • The paper quantifies the overthinking phenomenon in o1-like LLMs by introducing novel efficiency metrics that measure redundant computation after the first correct answer.
  • It employs MDL-based preference optimization techniques, including SFT, DPO, RPO, and SimPO, to train models to generate concise and correct reasoning paths.
  • Experiments reveal significant token reductions—over 50% in some cases—while largely maintaining or slightly improving reasoning accuracy on benchmark datasets.

This paper investigates the phenomenon of "overthinking" in LLMs exhibiting scaled inference-time computation, exemplified by systems analogous to OpenAI's O1 (termed "o1-like" models) (2412.21187). While these models achieve strong performance on complex reasoning tasks via extended Chain-of-Thought (CoT) generation and exploration of multiple solution strategies, they often allocate excessive computational resources to simple problems, resulting in significant inefficiency. The research introduces metrics to quantify this inefficiency and proposes mitigation strategies rooted in the Minimum Description Length (MDL) principle to enhance computational resource calibration during inference.

Problem Definition and Quantification of Overthinking

The core issue addressed is the tendency of o1-like models, such as QwQ-32B-Preview and DeepSeek-R1-Preview, to generate excessively long responses containing multiple, often redundant, solution attempts, irrespective of the input problem's intrinsic difficulty. This contrasts with conventional LLMs that typically produce a single, direct solution path. Empirical analysis across mathematical reasoning benchmarks (ASDIV, GSM8K, MATH500) reveals that these models generate, on average, 3.2-3.6 distinct solution attempts per query (Figure 3 in the paper), consuming substantially more tokens (~700-2400 tokens on average) compared to baselines like Llama-3.3-70B-Instruct (~150-600 tokens) (Table 1).

To formally quantify this inefficiency, two novel metrics suites are introduced:

  1. Outcome Efficiency (ξO\xi_O, ξO+\xi_O^+): These metrics assess the computational cost incurred after the first correct answer is generated within a response. ξO\xi_O is defined as the ratio of tokens in the shortest prefix yielding the first correct answer to the total tokens generated (Eq 1). ξO+\xi_O^+ is a stricter variant considering only valid responses (Eq 2). Low outcome efficiency indicates significant computation spent on redundant verification or exploration post-solution.
  2. Process Efficiency (ξP\xi_P, ξP+\xi_P^+): These metrics evaluate the diversity and conciseness of the reasoning process itself. They measure the ratio of tokens contributing to distinct solution perspectives relative to the total tokens generated (Eq 3, 4). Low process efficiency suggests repetitive or non-contributory reasoning steps within the generated CoT.

Calculations show o1-like models exhibit significantly lower efficiency scores compared to standard models. For instance, QwQ-32B-Preview displayed outcome efficiencies (ξO\xi_O) between 41-52% and process efficiencies (ξP\xi_P) between 66-72% on the tested datasets (Table 1), quantitatively confirming the overthinking behavior. Further analysis indicated diminishing returns, where later solution attempts provided marginal accuracy gains (Figure 4) and exhibited reduced diversity (Figure 5).

Mitigation Strategy via Minimum Description Length Principle

The proposed mitigation strategy draws inspiration from Occam's Razor and the Minimum Description Length (MDL) principle, suggesting that the most efficient (shortest) correct reasoning path should be preferred. The objective is formulated as finding the rationale rr that minimizes its length r||r|| subject to correctness, given an input xx and model parameters θ\theta:

r^=argminr{rrp(x;θ)}s.t. correctness(r,answer(x))\hat{r} = \arg\min_{r} \{ ||r|| \,|\, r \sim p(\cdot|x; \theta) \} \quad \text{s.t. correctness}(r, \text{answer}(x))

To operationalize this, a self-training paradigm is employed. The base o1-like model (QwQ-32B-Preview) generates multiple (K=10) candidate responses for problems from a suitable dataset (PRM12K). These responses are then used to construct training data for preference optimization, aiming to teach the model to favor efficient reasoning.

Two primary strategies for constructing preference pairs (rpreferred,rrejectedr_{preferred}, r_{rejected}) were explored:

  1. Response-Level Contrast: Pairs are formed by contrasting the shortest correct response generated for a problem against the longest correct response for the same problem. Statistical analysis of generated responses justified this approach (Figures 6, 7).
  2. Solution-Level Contrast: This approach, implemented via SimPOSolution_{Solution}, likely involves contrasting concise, correct individual solutions within a response against less efficient solutions or potentially the entire lengthy response containing them, encouraging optimality at a finer granularity.

These preference pairs are then used to fine-tune the base model using various techniques:

  • Supervised Fine-Tuning (SFT): Training directly on the preferred (shortest correct) responses.
  • Preference Optimization Algorithms: Utilizing methods like Direct Preference Optimization (DPO), Rank-based Preference Optimization (RPO), and Simple Preference Optimization (SimPO) with the constructed contrastive pairs.

Experimental Evaluation and Results

Experiments were conducted using QwQ-32B-Preview as the base model, fine-tuned with the proposed methods, and evaluated on ASDIV, GSM8K, MATH500, AIME, and GPQA datasets.

Key findings include:

  • Efficiency Improvements: Preference optimization methods, particularly SimPO, significantly reduced computational overhead. SimPOSolution_{Solution} demonstrated the most substantial reductions, decreasing token counts by over 50% and average solution rounds from 3.2 to 1.1 on GSM8K (Tables 2, 3). This effectively shifted the model's behavior towards generating single, concise solutions similar to conventional LLMs, while largely preserving accuracy on easier datasets.
  • Accuracy Maintenance: Most methods maintained or slightly improved accuracy compared to the baseline QwQ-32B-Preview, especially on simpler datasets (ASDIV, GSM8K). For instance, DPO and RPO methods applied at the response level slightly improved accuracy on GSM8K and MATH500 while reducing token counts (Table 2).
  • Trade-offs on Complex Tasks: The highly efficient SimPOSolution_{Solution} method exhibited a slight degradation in accuracy on the most challenging datasets (MATH500, AIME), suggesting a potential trade-off where excessive pruning of the search space might hinder performance on problems genuinely requiring extensive exploration.
  • Reduction in Redundant Computation: The methods successfully reduced "Additional Response" computation (tokens/rounds generated after the first correct answer), indicating improved calibration between computational effort and task completion (Tables 2, 3).
  • State-of-the-Art Claims: The paper reports achieving state-of-the-art performance on MATH500 (value placeholder xx% used) and improved results on GPQA and AIME relative to other open-source models (Table 5), attributing this partly to the mitigation of overthinking and associated phenomena like inversion errors (correct answers being erroneously revised).

Implications for Computational Efficiency

The research carries significant implications for the practical deployment and operational cost of advanced reasoning LLMs:

  1. Reduced Inference Costs: Mitigating overthinking directly translates to lower computational expenditure per query, as fewer tokens are generated. This reduction (potentially >50% with methods like SimPOSolution_{Solution}) lowers the operational costs associated with GPU time and energy consumption.
  2. Lower Inference Latency: Generating fewer tokens reduces the time required for inference, improving response times and making models more suitable for interactive or real-time applications.
  3. Increased System Throughput: By decreasing the computational load per request, serving infrastructures can handle a higher volume of concurrent requests, enhancing overall system throughput and user capacity.
  4. Efficient Resource Allocation: Models trained to avoid overthinking on simpler tasks free up valuable computational resources. These resources can then be directed towards handling more users or allocated to tackle genuinely complex problems that benefit from extended computational budgets.
  5. Refined Scaling Perspectives: The findings challenge the notion that simply increasing inference-time computation universally improves reasoning. It underscores the necessity of calibrated computation, aligning resource allocation with task complexity, potentially influencing future designs for adaptive computation mechanisms in LLMs.

Conclusion

This work provides a systematic analysis of the "overthinking" inefficiency observed in o1-like LLMs, characterized by excessive computational effort on simple tasks (2412.21187). By introducing novel efficiency metrics (ξO,ξP\xi_O, \xi_P) and leveraging the MDL principle via preference optimization techniques (SFT, DPO, RPO, SimPO) trained on self-generated contrastive data, the authors demonstrate effective mitigation strategies. These methods significantly reduce computational overhead (token count, solution rounds) while largely maintaining or even improving reasoning accuracy, particularly on less complex benchmarks, contributing towards more computationally efficient and practical deployment of advanced reasoning models.

Youtube Logo Streamline Icon: https://streamlinehq.com

HackerNews

  1. The Overthinking of O1-Like LLMs (3 points, 1 comment)
Reddit Logo Streamline Icon: https://streamlinehq.com