Papers
Topics
Authors
Recent
2000 character limit reached

Autoregressive & Masked-Diffusion Training

Updated 18 December 2025
  • Autoregressive and masked-diffusion training are two paradigms for generative modeling, where AR uses strict sequential prediction and MD employs stochastic masking for global context.
  • AR models optimize next-token prediction with local dependencies while MD models use randomized mask schedules for implicit data augmentation and improved generalization.
  • Recent research shows that hybrid approaches combining AR and MD can match benchmark performance in language, vision, and recommendation tasks.

Autoregressive and masked-diffusion training define two principal paradigms for probabilistic modeling and generative learning in discrete domains, including language, vision, and recommendation. While both frameworks leverage deep neural architectures—primarily transformers—their inductive biases, training objectives, and sampling algorithms introduce complementary strengths and limitations. Recent research advances demonstrate not only competitive performance between the paradigms under specific settings, but also deep connections: masked diffusion models admit autoregressive decompositions over learned or randomized generation orders, and, under suitable loss reweightings and mask schedules, can replicate or outperform autoregressive baselines in data-constrained and compositional tasks.

1. Formulations: Factorizations, Objectives, and Architectures

Autoregressive (AR) models impose a strict left-to-right (or otherwise ordered) factorization of the joint data distribution over sequences. For a sequence x1:Lx_{1:L},

pAR(x1:L)=j=1Lpθ(xjx<j),p_{\mathrm{AR}}(x_{1:L}) = \prod_{j=1}^{L} p_\theta(x_j \mid x_{<j}),

with θ\theta denoting model parameters, usually implemented as a causal transformer. The canonical training loss is next-token cross-entropy,

LAR(θ)=j=1Llogpθ(xjx<j),\mathcal{L}_{\mathrm{AR}}(\theta) = -\sum_{j=1}^{L} \log p_\theta(x_j \mid x_{<j}),

optimized under full supervision (“teacher forcing”).

Masked-diffusion (MD) models instead define a forward stochastic “noising” process q(xtx0)q(x_t \mid x_0) over a masking schedule, progressively corrupting clean data x0x_0 to a fully masked variant xtx_t. The backward (“denoising”) model pθ(x0xt)p_\theta(x_0 \mid x_t) learns to invert this process in probabilistic or deterministic steps. The continuous-time loss reduces, under mean parameterization, to a weighted integral of per-token cross-entropy,

LMD(θ)=01w(t) Ex0,xtq(x0,t)[i:xti=[MASK]logpθ(x0ixt,t)]dt,\mathcal{L}_{\mathrm{MD}}(\theta) = \int_0^1 w(t)~\mathbb{E}_{x_0, x_t \sim q(\cdot \mid x_0, t)} \left[ -\sum_{i: x_t^i = [\mathrm{MASK}]}\log p_\theta(x_0^i \mid x_t, t)\right] dt,

where w(t)w(t) is a time-dependent weighting derived from the ELBO and mask schedule (Shi et al., 6 Jun 2024, Sahoo et al., 11 Jun 2024).

AR models usually are restricted to causal, decoder-style transformers; masked-diffusion models employ encoder-only, bidirectional attention, enabling conditioning on arbitrary observed/masked token patterns at each training step.

2. Training Dynamics and Inductive Bias

Autoregressive training exposes the model to a sequence of deterministic prediction tasks, each determined by a unique context of preceding tokens. This results in efficient likelihood maximization but strongly conditions parameter updates on local dependencies, limiting exposure to “global” context and inducing the so-called “reversal curse” in knowledge injection and QA (Pan et al., 10 Oct 2025). AR models tend to saturate rapidly under repeated passes over limited data, exhibiting a data half-life RD31R_D^* \approx 31 (epochs), and overfit or plateau in validation loss at low data regimes (Prabhudesai et al., 21 Jul 2025).

Masked-diffusion training, by contrast, repeatedly perturbs each sequence under randomized masking ratios and locations, thereby exposing the model to an exponentially larger set of partial contexts and prediction subproblems. This acts as implicit data augmentation: the effective data reuse half-life reaches RD494R_D^*\approx 494, preventing overfitting and supporting improved generalization in data-constrained settings (Prabhudesai et al., 21 Jul 2025). On downstream tasks, masked-diffusion models outperform AR models as soon as the available compute exceeds a critical threshold scaling as Ccrit(U)191U2.174C_{\rm crit}(U) \approx 191 U^{2.174}, with UU the number of unique tokens (Prabhudesai et al., 21 Jul 2025).

3. Forward and Reverse Processes: Masked-Diffusion Mechanism

The forward process in masked-diffusion applies a sequence of discrete masking operations parameterized by a schedule αt\alpha_t (cumulative “keep” probability): q(xtx0)=Cat(xt;αtx0+(1αt)m),q(x_t \mid x_0) = \mathrm{Cat}(x_t; \alpha_t x_0 + (1-\alpha_t)m), where mm denotes the “absorbing” [MASK] state (Shi et al., 6 Jun 2024). State-dependent and coordinate-specific masking schedules α(t)\alpha_\ell(t) generalize this to enable learned or structured unmasking orders (Garg et al., 24 Nov 2025). At each time, otherwise independent tokens are masked stochastically, yielding a partially observed xtx_t.

Reverse denoising is executed either via iterative refinement (progressively unmasking according to a schedule), semi-autoregressive blockwise strategies, or in the extreme, full parallel decoding. The denoising model predicts each masked token as a conditional categorical distribution given the observed tokens and the current mask, potentially decomposable as

pθ(x0xt)=iMpθ(x0ixt,t),p_\theta(x_0 \mid x_t) = \prod_{i \in \mathcal M} p_\theta(x_0^i \mid x_t, t),

with M\mathcal M the set of masked indices at step tt (Shi et al., 6 Jun 2024, Sun et al., 29 Sep 2025).

4. Connections, Limitations, and Order Learning

Masked diffusion models are formally mixtures of autoregressive models over random or learned generation orders. Under a multivariate mask schedule α\alpha_\ell, the continuous-time ELBO decomposes as

L(θ)=πP(π)Ex0[LLπ(x0;θ)] ,\mathcal{L}(\theta) = - \sum_{\pi} P(\pi) \mathbb{E}_{x_0}[LL_{\pi}(x_0; \theta)]~,

where each π\pi is a permutation of token positions (a decoding order), LLπLL_\pi is the AR log-likelihood along π\pi, and P(π)P(\pi) is the probability of sampling π\pi under the joint schedule (Garg et al., 24 Nov 2025). The model thus “secretly” learns over random or explicitly optimized AR orders, interpolating between pure AR (single order) and BERT-like masking (uniform order).

However, standard masked diffusion models encode only the marginals over masked positions, not the joint (Sun et al., 29 Sep 2025). As a result, parallel unmasking of many tokens can yield incoherent text beyond a small subset (typically n6n \le 6), and the best practical decoders are semi-autoregressive (blockwise) or confidence-adaptive single-token decoders, both of which reduce the joint-perplexity gap to AR but do not close it entirely (Sun et al., 29 Sep 2025).

5. Empirical Benchmarks and Applications

Empirical comparisons indicate that masked-diffusion models trained with appropriate scheduling, loss weighting, and engineering practices can approach or even match AR models on both perplexity and downstream tasks. State-of-the-art masked-diffusion LMs achieve perplexity competitive with AR on large benchmarks such as OpenWebText, with observed reductions in overfitting and improved robustness in few-shot and backward-inference settings (Sahoo et al., 11 Jun 2024, Shi et al., 6 Jun 2024, Prabhudesai et al., 21 Jul 2025, Pan et al., 10 Oct 2025).

In generative recommendation, masked-diffusion offers substantial gains in data efficiency and “coarse-grained” recall, with the ability to predict multiple semantic-ID tokens per item in parallel, outperforming AR approaches especially at high retrieval depths and under data scarcity (Shah et al., 28 Nov 2025). For vision–language learning, masked-diffusion captioning equalizes the visual supervision gradient across positions, matching or exceeding autoregressive baselines in vision transfer and compositionality (Feng et al., 30 Oct 2025).

6. Practical Considerations and Training Enhancements

Key obstacles for masked-diffusion training arise from its higher intrinsic variance: three dominant sources appear (A) masking pattern noise, (B) masking rate noise, (C) data noise, as opposed to (C) alone in ARMs (Jia et al., 22 Nov 2025). Pareto-optimal time samplers (P-POTS) and mirrored-mask variance reduction (MIRROR) demonstrably control (A)+(B), improving both average accuracy (+7–8 points on reasoning tasks) and run-to-run convergence reliability—essential for stable model scaling (Jia et al., 22 Nov 2025). Blockwise reverse-order training and semi-AR generation align training and inference distributions, further reducing quality dropoff at high parallelism rates (Sun et al., 29 Sep 2025, Yang et al., 28 Sep 2025).

Aligning training and inference schedules is critical: masked-diffusion models trained purely on random masks perform suboptimally when inference proceeds via deterministic progressive unmasking; reinforcement learning approaches (e.g., MDPO, CJ-GRPO) and trajectory-consistent policy optimization (MDPO) align these processes, yielding an order-of-magnitude reduction in sample and optimization inefficiency (He et al., 18 Aug 2025, Yang et al., 28 Sep 2025).

7. Hybridization and Theoretical Unification

Recent work unifies AR and masked-diffusion paradigms under frameworks such as Autoregressive Diffusion Models (ARDMs), which generalize order-agnostic AR models and absorbing diffusion through strict schedule control, supporting both sequential and parallel generation with efficiency–quality trade-offs tunable via mask scheduling and dynamic programming (Hoogeboom et al., 2021). Hybrid approaches such as DC-AR combine masked-AR for structural prediction with lightweight diffusion refinement for high-fidelity image generation, leveraging the speed of AR and the expressiveness of diffusion in a two-stage pipeline (Wu et al., 7 Jul 2025). Masked-diffusion finetuning recipes augment AR models’ data efficiency for challenging knowledge-injection tasks, even closing the “reversal curse” with masked fine-tuning objectives (Pan et al., 10 Oct 2025).


These developments suggest that masked-diffusion training, via judicious engineering of loss, scheduling, and decoding algorithms, can supplement or replace conventional autoregressive maximum-likelihood in settings where data efficiency, parallel decoding, global context, or flexible conditional generation are paramount, while careful alignment of train/infer policies and variance control are necessary for full realization of its theoretical advantages.

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Autoregressive and Masked-Diffusion Training.