Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 91 TPS
Gemini 2.5 Pro 55 TPS Pro
GPT-5 Medium 40 TPS
GPT-5 High 40 TPS Pro
GPT-4o 94 TPS
GPT OSS 120B 477 TPS Pro
Kimi K2 231 TPS Pro
2000 character limit reached

Diffusion LLMs Can Do Faster-Than-AR Inference via Discrete Diffusion Forcing (2508.09192v1)

Published 8 Aug 2025 in cs.LG and cs.AI

Abstract: Diffusion LLMs (dLLMs) have emerged as a promising alternative to autoregressive (AR) LLMs for text generation, with the potential to decode multiple tokens in a single iteration. However, none of the existing open-source dLLMs have achieved superior inference speed over AR LLMs of similar size. This paper breaks this barrier based on a simple and effective strategy named discrete diffusion forcing (D2F). D2F equips dLLMs with two key capabilities: (1) block-wise autoregressive generation to enable KV cache utilization; (2) prediction of following tokens without requiring completion of prior blocks for inter-block parallel decoding. In this way, the vanilla dLLMs are refurbished into an AR-diffusion hybrid paradigm for efficient inference. D2F can be implemented with an asymmetric distillation process based on pre-trained dLLMs. We further propose a pipelined parallel decoding algorithm, which enables a trade-off between efficiency and efficacy. Empirically, D2F dLLMs achieve more than $\mathbf{2.5\times}$ inference speed than LLaMA3 and Qwen2.5 on GSM8K. Compared to vanilla dLLMs like LLaDA and Dream, the acceleration can be more than $\mathbf{50\times}$ while maintaining comparable output quality. The code is available at https://github.com/zhijie-group/Discrete-Diffusion-Forcing.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper introduces a novel AR-diffusion hybrid (D2F) that uses block-wise causal attention and inter-block parallel decoding to surpass AR inference speeds.
  • Experimental results show up to 2.5× speedup over autoregressive LLMs and over 50× acceleration compared to vanilla diffusion models with competitive quality.
  • Ablation studies reveal that optimal block sizes and pipeline hyperparameters are critical for balancing throughput and output accuracy.

Discrete Diffusion Forcing Enables Faster-Than-Autoregressive Inference in Diffusion LLMs

Introduction

The paper "Diffusion LLMs Can Do Faster-Than-AR Inference via Discrete Diffusion Forcing" (2508.09192) presents Discrete Diffusion Forcing (D2F), a training and inference paradigm for Diffusion LLMs (dLLMs) that achieves inference speeds surpassing autoregressive (AR) LLMs of comparable scale. D2F introduces a hybrid AR-diffusion generation scheme, leveraging block-wise causal attention and inter-block parallel decoding, which enables efficient KV cache utilization and aggressive parallel token generation. This work establishes the first open-source dLLMs that outperform AR LLMs in throughput while maintaining competitive output quality.

Background: Diffusion LLMs and Inference Bottlenecks

Diffusion LLMs generate text by iteratively denoising a masked sequence, allowing for parallel prediction of all tokens at each step. Theoretically, this paradigm offers substantial acceleration over AR models, which decode tokens sequentially. However, practical dLLMs have been hampered by two major bottlenecks:

  • Bidirectional Attention: Standard dLLMs use bidirectional attention, which is incompatible with KV cache, resulting in redundant computation across denoising steps.
  • Conditional Independence: Parallel decoding assumes conditional independence among tokens, which is violated in natural language, necessitating more iterative steps for high-quality outputs.

Previous acceleration methods, such as block-wise sequential generation and approximate KV cache, have not simultaneously achieved precise KV cache and efficient parallel decoding. Consequently, open-source dLLMs have lagged behind AR models in inference speed.

Discrete Diffusion Forcing: Methodology

D2F restructures dLLM generation into a block-wise AR-diffusion hybrid, enabling both KV cache compatibility and inter-block parallelism. The key components are:

  • Block-wise Causal Attention: The answer sequence is partitioned into blocks, each with a monotonically increasing masking ratio. Attention is causal across blocks and bidirectional within blocks, allowing KV cache for completed blocks.
  • Inter-block Parallel Decoding: The model is trained to predict future blocks conditioned on partially denoised predecessors, enabling parallel decoding without waiting for full completion of prior blocks. Figure 1

    Figure 1: Overview of D2F training: answers are divided into blocks with increasing masking ratios, and the model is trained to mimic a bidirectional teacher conditioned on partially denoised preceding tokens.

Asymmetric Distillation

D2F dLLMs are distilled from pre-trained bidirectional dLLMs using an asymmetric KL divergence loss. The teacher predicts each block with a global view of all noisy blocks, while the student (D2F dLLM) predicts using only a causally restricted view. The student architecture differs solely in its attention mask, enforcing block-wise causality.

Pipelined Parallel Decoding

During inference, D2F maintains a pipeline of active blocks. A new block is added when the completion ratio of the last block exceeds a threshold τadd\tau_{add}. Newly added blocks are semi-activated for conservative decoding and become fully activated when their predecessor reaches τact\tau_{act}, allowing aggressive parallel decoding. Figure 2

Figure 2: Pipelined parallel decoding: blocks are decoded in parallel, with dynamic addition and activation based on completion thresholds.

Experimental Results

D2F is evaluated on LLaDA-Instruct-8B and Dream-Base-7B, distilled on the Bespoke-Stratos-17k dataset. Benchmarks include GSM8K, MATH, HumanEval, and MBPP. D2F achieves:

  • Up to 2.5× speedup over AR LLMs (LLaMA3-Instruct-8B, Qwen2.5-Base-7B) on GSM8K.
  • Over 50× acceleration compared to vanilla dLLMs (LLaDA, Dream) with comparable output quality.
  • Superior throughput-performance trade-off: D2F maintains output quality at high throughput, whereas vanilla dLLMs suffer substantial degradation when decoding steps are reduced. Figure 3

    Figure 3: D2F dLLMs surpass AR LLMs in inference speed for up to 2.5×2.5\times.

    Figure 4

    Figure 4: Throughput vs. performance trade-off: D2F achieves a more favorable balance compared to vanilla dLLMs.

Ablation Studies and Analysis

Ablation studies dissect the contributions of block size, pipeline hyperparameters, and noise scheduling:

  • Block Size: Increasing block size initially improves performance but eventually degrades throughput and output quality. Optimal block size selection is critical for balancing speed and accuracy. Figure 5

    Figure 5: Ablation paper on block size during inference: performance peaks at moderate block sizes, with throughput decreasing as block size increases.

  • Pipeline Hyperparameters: Dual-state decoding (semi-activated then fully activated blocks) outperforms single-state pipelines, yielding higher scores and throughput.
  • Noise Scheduling: Structured progressive masking (D2F) outperforms random masking schedules, substantiating the efficacy of the D2F training objective.

Implications and Future Directions

D2F demonstrates that dLLMs can achieve faster-than-AR inference without sacrificing output quality, challenging the prevailing assumption that AR models are inherently superior in efficiency. The AR-diffusion hybrid paradigm introduced by D2F is extensible to other discrete generative tasks, including code synthesis and mathematical reasoning. The pipelined parallel decoding strategy and asymmetric distillation framework are broadly applicable to other sequence modeling domains.

Theoretical implications include the principled extension of diffusion forcing from continuous to discrete data, bridging the gap between AR and diffusion-based generative models. Practically, D2F enables deployment of dLLMs in latency-sensitive applications, such as real-time dialogue systems and large-scale batch generation.

Future research may explore:

  • Scaling D2F to larger models and longer contexts
  • Integration with advanced sampling and confidence-based remasking strategies
  • Application to multimodal and multilingual generative tasks
  • Further optimization of block partitioning and pipeline scheduling

Conclusion

Discrete Diffusion Forcing (D2F) establishes a new regime for dLLMs, enabling faster-than-AR inference through block-wise causal attention and inter-block parallel decoding. Empirical results validate substantial speedups and competitive output quality, positioning D2F as a foundational technique for efficient discrete sequence generation. The AR-diffusion hybrid paradigm and pipelined parallel decoding introduced by D2F have significant implications for the future of generative modeling in both research and production environments.

HackerNews