Stable injection of residual context into diffusion LLM input embeddings
Develop a method to inject soft-token residual context into the input embeddings of diffusion large language models that preserves the discrete masking scheme and avoids unstable recursive dependencies during training.
References
However, naively adding the soft tokens to input embeddings disrupts the discrete masking scheme dLLMs rely on and creates unstable recursive dependencies during training, leaving injecting residual context into dLLMs an open question for RCD to solve.
— Residual Context Diffusion Language Models
(2601.22954 - Hu et al., 30 Jan 2026) in Section 2.2 (Soft Tokens)