Extent of Chain-of-Thought Faithfulness to Underlying Computation
Determine the degree to which chain-of-thought outputs generated by large language models faithfully represent the models' underlying computational processes that produce their final answers.
References
While current reasoning LLMs produce chain-of-thought (CoT) outputs that ostensibly reflect their reasoning, it remains unclear to what extent these outputs faithfully represent the model's true computational process \citep{lanham2023measuring, barez2025chain, chen2025reasoning}.
— Balancing Faithfulness and Performance in Reasoning via Multi-Listener Soft Execution
(2602.16154 - Sivakumaran et al., 18 Feb 2026) in Section 1 (Introduction), first paragraph