Whether LLMs Genuinely Perform Latent Reasoning
Determine whether large language models internally perform latent multi-hop reasoning—retrieving intermediate bridge entities and propagating inference—rather than simply memorizing multi-hop question patterns as atomic facts.
References
The inner workings of LLMs are themselves a controversial topic, and some unconfirmed issues about these mechanisms are fatal to the task of knowledge updating: one such question is whether LLMs genuinely perform latent reasoning.
— Open Problems and a Hypothetical Path Forward in LLM Knowledge Paradigms
(2504.06823 - Ye et al., 9 Apr 2025) in Section 3.1 (Challenges in Updating LLM Knowledge)