Conjecture: Hash-conditioning enables latent plan-conditioned generation
Establish whether hash-conditioning in Transformers—training with random hash string prefixes and using novel hash prefixes at inference—induces a latent-plan-conditioned generation process in which a single random "leap of thought" z is selected before decoding and tokens are generated from the conditional distribution p(s | z), rather than by marginalizing over multiple latent plans via output-layer temperature sampling.
References
We conjecture that hash-conditioning enables this conditioned token generation.
— Roll the dice & look before you leap: Going beyond the creative limits of next-token prediction
(2504.15266 - Nagarajan et al., 21 Apr 2025) in Section 1 (Introduction)