Dice Question Streamline Icon: https://streamlinehq.com

Does the lack of frontier expansion with increased NFE persist at larger scales?

Ascertain whether the observed absence of entropy–perplexity frontier expansion as the number of function evaluations (NFE) increases for discrete diffusion language models is merely an artifact of the model scale used in the experiments, or whether this behavior persists for larger diffusion models.

Information Square Streamline Icon: https://streamlinehq.com

Background

In extended OWT experiments, the authors observe that increasing NFE does not significantly expand the entropy–perplexity frontier; temperature tuning appears to recover most benefits that additional steps would otherwise provide.

They explicitly state uncertainty about whether this phenomenon stems from the relatively small scale of their models or represents a general property that would also hold at larger scales.

References

We observe that the frontiers are largely the same, implying that the performance improvements obtained by increasing NFE can be achieved via temperature tuning. We do not know whether this is simply an artifact of this specific scale of models, or whether this is a persistent phenomenon for larger diffusion models.

CANDI: Hybrid Discrete-Continuous Diffusion Models (2510.22510 - Pynadath et al., 26 Oct 2025) in Appendix, Section 6: OWT Text Generation, Frontier Curves for Higher NFE