Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 25 tok/s Pro
GPT-5 High 28 tok/s Pro
GPT-4o 86 tok/s Pro
Kimi K2 203 tok/s Pro
GPT OSS 120B 445 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Degeneration-of-Thought Problem

Updated 24 October 2025
  • Degeneration-of-Thought is the decline in fidelity, coherence, and diversity of internal reasoning steps in neural models, leading to accumulated errors in complex tasks.
  • Innovative methods such as skip connections, tree-based and DAG-based reasoning frameworks actively mitigate information loss and redundancy in multi-step inference.
  • Information-theoretic insights, including Fisher Information decay, drive architectural innovations like external CoT guidance and diffusion techniques to address DoT.

The Degeneration-of-Thought (DoT) problem refers to the decline in fidelity, coherence, diversity, or structure of intermediate cognitive or reasoning steps in deep neural architectures and LLMs as they process complex tasks. The phenomenon manifests in forms such as loss of informativeness in latent variables, repetitive and redundant output, shallow reasoning due to premature switching, or accumulation of errors along lengthy multi-step inference chains. Several recent frameworks and architectural innovations have been introduced to diagnose, quantify, and mitigate DoT in both generative and reasoning models.

1. Core Definition and Phenomena

Degeneration-of-Thought is observed when neural models—specifically deep feed-forward architectures (e.g., VAEs, LLMs with multi-step reasoning interfaces)—incrementally lose critical information throughout the processing pipeline. Key symptoms include:

These diverse manifestations share a common root: as models traverse increasingly deep, complex, or multi-stage reasoning chains, information becomes diluted, errors may accumulate, diversity may collapse, and the model's ability to maintain robust, meaningful thought trajectories degenerates.

2. Information-Theoretic Foundations: Fisher Information Loss

In deep VAEs, DoT is directly induced by the decay of Fisher Information at each layer: IF(ϕl+1)=IF(ϕl)(ϕlF(X)ϕl+1F(X))2\mathcal{I}_F(\phi_{l+1}) = \mathcal{I}_F(\phi_l) \cdot \left(\frac{\nabla_{\phi_l} F(X)}{\nabla_{\phi_{l+1}} F(X)} \right)^2 This recursion implies IF(ϕl+1)IF(ϕl)\mathcal{I}_F(\phi_{l+1}) \leq \mathcal{I}_F(\phi_l), formalizing the unavoidable loss of parameter information with each layer's propagation (Zheng et al., 2018). In VAEs, this leads to the three classical degeneration scenarios: latent codes lose connection to inputs, reconstructions degrade in quality, and the network's representation power wanes with depth.

To mitigate this, skip connections (as in SCVAE) supplement the information flow: I(hl,c(hlk))(ϕl)=Ihl(ϕl)+Ic(hlk)hl(ϕl)\mathcal{I}_{(h_l, c(h_{l-k}))}(\phi_l) = \mathcal{I}_{h_l}(\phi_l) + \mathcal{I}_{c(h_{l-k})|h_l}(\phi_l) which ensures additional preservation of Fisher Information across layers.

3. Mechanisms in Reasoning Models: Trees, DAGs, and Diffusion

For large reasoning models (LRMs), DoT is addressed via structural reforms in the modeling of cognitive steps (“thoughts”):

  • Tree of Thoughts (ToT): The reasoning process is modeled as a tree, with each node being a coherent intermediate step. Instead of a strict linear chain, multiple candidate thoughts are generated, evaluated, and the search algorithms (BFS, DFS) allow for strategic lookahead, backtracking, and global planning (Yao et al., 2023). This tree-based exploration mitigates DoT by systematically comparing, pruning, and revisiting branches.
  • Diagram of Thought (DoT): The iterative reasoning steps form a directed acyclic graph (DAG); the process cycles between proposing, critiquing, refining, and finally synthesizing via a formal colimit operation in Topos Theory: Final Conclusion=colim D\text{Final Conclusion} = \mathrm{colim}\ D with each reasoning node aggregated for logical consistency and robustness (Zhang et al., 16 Sep 2024).
  • Diffusion-of-Thought (DoT): Chain-of-thought reasoning in diffusion models leverages a latent denoising process. Reasoning steps diffuse temporally, enabling global revisiting and intrinsic self-correction since each step can be re-refined throughout the denoising trajectory. Scheduled sampling during training further enhances the model's ability to correct its own mistakes (Ye et al., 12 Feb 2024).

4. Diversity, Underthinking, and Overthinking Control

Degeneration arises not only from loss of information but also from reasoning inefficiencies:

  • Underthinking: In models like o1-Like LLMs, frequent thought switching leads to shallow exploration; token efficiency drops sharply for incorrect responses. The underthinking metric (ξUT\xi_{UT}) quantifies wasted reasoning as

ξUT=1Ni=1N(1T^iTi)\xi_{UT} = \frac{1}{N} \sum_{i=1}^N \left(1 - \frac{\hat{T}_i}{T_i}\right)

TIP (Thought Switching Penalty) applies a logit penalty to discourage premature transitions, increasing solution accuracy without fine-tuning (Wang et al., 30 Jan 2025).

  • Overthinking: Slow-thinking models may overgenerate tokens for simple tasks. DAST introduces a Token Length Budget (TLB) to adaptively penalize superfluous reasoning for easy queries while encouraging sufficient depth for complex ones (Shen et al., 6 Mar 2025).
  • External Thought Manipulation: ThoughtMani proposes inserting externally-generated CoTs—delimited by > , —to forcibly short-circuit redundant internal reasoning; this yields 18–37% token reduction and improved safety alignment in experiments (Liu et al., 18 Apr 2025).

5. Architectural Innovations for Mitigation

Several recent architectures generalize and formalize strategies for combating DoT:

  • Everything of Thoughts (XoT): Unifies performance, efficiency, and flexibility by using a pretrained policy-value network with MCTS for generating candidate thought trajectories, then collaborating with an LLM for revision, thus keeping LLM calls minimal and diversity high (Ding et al., 2023).
  • Tree of Problems (ToP): Decomposes complex tasks into trees of identical subtasks, using CoT at each leaf and merging solutions bottom-up; this modular composition significantly reduces error accumulation and preserves structured reasoning (Zebaze et al., 9 Oct 2024).
  • Dual Engines of Thoughts (DEoT): For open-ended analysis, a Breadth Engine (to foster diversity) and Depth Engine (to deepen investigation) are orchestrated for comprehensive coverage, actively breaking cycles of stagnation and shallow reasoning (Yu et al., 10 Apr 2025).
  • Lateral Tree-of-Thoughts (LToT): Separates utility from logical consistency, maintains a dual frontier (mainlines for exploitation, laterals for cheap broad exploration), and allocates compute in capped rungs for diversity. The lateral exploration cost is pseudolinear: Θ(N0logηN0)\Theta(N_0 \log_\eta N_0) making large test-time breadth feasible without exponential inefficiency (Madahar, 1 Oct 2025).

6. Data Perspective and Training Signal

DoT in text generation is fundamentally correlated with repetition in training data. Experiments partitioning datasets by n-gram repetition reveal amplification of repetition in outputs by trained models. Methods such as repetition dropout—modifying the attention mask to ignore over-repeated patterns—minimize degeneracy in generated text. Penalizing repetitions unifies previous effective approaches including high-inflow word penalties, SCALEGRAD, and objective adaptation (Li et al., 2023).

7. Theoretical Barriers and Future Directions

A critical theoretical insight is that transformer-based LLMs—being deterministic functions of input—cannot form genuine internal "thoughts" in the feature space: I(Statet;YtX1:t)=0I(\text{State}_t; Y_t | X_{1:t}) = 0 To enable authentic thought processes, architectural modifications such as injection of random instance-specific vectors, recurrence, and instance-level training must be incorporated (Jahrens et al., 12 Mar 2025). These modifications could fundamentally improve planning, consistency, and argument formation by enabling the feature space to perform internal decision-making prior to token sampling.


The Degeneration-of-Thought problem encompasses a range of phenomena across deep generative and reasoning architectures. Its diagnosis and remediation require careful attention to information preservation, structure, diversity, and the underlying training signals. Recent advances across architectural, inference-time, and data-centric axes provide multiple technically rigorous strategies for detection and mitigation, with ongoing research dedicated to establishing both practical metrics and theoretical guarantees.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Degeneration-of-Thought (DoT) Problem.