Dice Question Streamline Icon: https://streamlinehq.com

Precise natural-language control of AutoDeco decoding

Determine whether joint training of the base transformer and the AutoDeco temperature and top-p heads enables precise, absolute control of token-level decoding parameters via natural-language commands (such as directives for low diversity or no randomness), and characterize the mechanism underlying AutoDeco’s emergent prompt-driven adjustments to predicted temperature and top-p values.

Information Square Streamline Icon: https://streamlinehq.com

Background

The paper reports an emergent capability in AutoDeco: the model can interpret high-level natural-language commands about creativity or certainty and accordingly adjust its predicted temperature and top-p on a token-by-token basis. While targeted training with a ranking loss improves the consistency of these adjustments, the authors observe that the resulting control is not yet precise or absolute.

They explicitly state that they lack a conclusive understanding of the phenomenon and hypothesize that achieving fine-grained control may require joint training of the base LLM and the AutoDeco heads. The released models therefore omit this experimental control feature pending further investigation.

References

However, we do not yet have a conclusive understanding of this phenomenon. While the trained model learns the correct directional adjustments, it does not achieve precise, absolute control. We hypothesize that achieving such fine-grained control may require joint training of the base LLM and the AutoDeco heads.

The End of Manual Decoding: Towards Truly End-to-End Language Models (2510.26697 - Wang et al., 30 Oct 2025) in Section “Emergent Control of Decoding via Natural Language”