Continuous Autoregressive Language Models
- Continuous Autoregressive Language Models (CALM) are defined as models that predict continuous latent vectors instead of discrete tokens, boosting semantic bandwidth and generation efficiency.
- They leverage high-fidelity autoencoding to compress token chunks with over 99.9% reconstruction accuracy, ensuring robust and error-tolerant latent representations.
- CALM employs likelihood-free training with energy score losses and BrierLM metrics, outperforming traditional autoregressive models in efficiency and scalability across modalities.
Continuous Autoregressive LLMs (CALM) are a paradigm in language modeling wherein the generative process transitions from discrete, token-by-token prediction to rich, continuous next-vector prediction. This shift enables dramatic improvements in semantic bandwidth, computational efficiency, and flexibility in generative modeling. CALM encompasses architectures and methodologies for text, audio, video, and cross-modal modeling, supported by specialized training, evaluation, and sampling tools. The framework incorporates concepts from information geometry, spectral analysis, and advanced autoencoding, pairing theoretical depth with empirical performance. The following sections delineate the foundational principles, architectures, mathematical formulations, and implications of CALM.
1. From Discrete Token Prediction to Continuous Next-Vector Modeling
Traditional autoregressive LLMs are constructed as sequential predictors over a discrete vocabulary, maximizing per token step. CALM reconceptualizes language as a sequence of continuous latent vectors, each encoding tokens, thereby amplifying the semantic payload per step and reducing the length of the generative sequence by a factor of %%%%2%%%% (Shao et al., 31 Oct 2025). Autoencoding is employed to compress tokens into , such that the sequence is represented as , with . The autoregressive model then operates on the latent space:
The autoencoder is trained to reconstruct each -chunk with high fidelity ( accuracy for practical chunk sizes), using cross-entropy loss over the decoded token sequence. Variational regularization ensures latent robustness, with the encoder outputting Gaussian parameters and KL regularization to the standard normal prior.
A similar shift to continuous autoregression characterizes recent models in audio (Simon et al., 8 Sep 2025), video (Yu et al., 17 Jun 2025), and flexible flows (Zhang et al., 1 Jul 2025). In each domain, modeling proceeds in continuous latent space, replacing quantization with high-fidelity compression and enabling rich, bandwidth-efficient generative steps.
2. High-Fidelity Autoencoding and Robust Latent Representations
The efficacy of CALM depends on constructing autoencoders that are both high-fidelity and robust to generative error accumulation. For text, an encoder (embeddings FFN linear projection) compresses a -token chunk; the decoder inverts this map, reconstructing tokens from the latent vector. The total loss is
KL clipping and dropout (masking latent and input components) are employed to ensure redundancy and error tolerance. In practice, moderate latent dimensionality and chunk sizes suffice for near-perfect reconstruction. The stochastic encoding of latents (sampling from ) ensures robustness under noise, facilitating stable autoregressive generation in the continuous domain.
In audio, a VAE-GAN backbone encodes the waveform as a sequence of continuous latents, optimized with joint time-domain, frequency-domain, adversarial, and KL losses (Simon et al., 8 Sep 2025). For video, continuous tokens extracted from pretrained VAEs are autoregressed with masked parallelism and curriculum learning (Yu et al., 17 Jun 2025). Across modalities, the elimination of lossy quantization is central to CALM’s fidelity and efficiency.
3. Likelihood-Free Training, Evaluation, and Sampling
Autoregressive modeling in continuous space invalidates the softmax-cross-entropy routine. CALM introduces likelihood-free training via strictly proper scoring rules—most notably, the energy score:
Empirical energy loss averages distances between model samples and data samples from the autoencoder posterior. Training the generative head (Energy Transformer) to minimize energy score ensures alignment of model and data distributions in latent space, without explicit likelihoods.
For evaluation, CALM employs BrierLM—a sample-based adaptation of the Brier score for n-gram language modeling. BrierLM is calculated as the geometric mean of Brier-1 to Brier-4 scores over sampled outputs:
This metric is strongly correlated with cross-entropy (Spearman’s rank: ), validating its use as a likelihood-free counterpart to perplexity.
Temperature sampling in continuous models—absent logits—uses a rejection sampling scheme, where samples per step yield for temperature , and fractional schemes generalize this process. This provides a controlled accuracy-diversity trade-off matching traditional temperature scaling.
4. Continuous Autoregressive Modeling Across Modalities and Architectures
CALM's principles have been generalized beyond text:
- Continuous Audio LLMs: Sequence modeling is performed in audio VAE latent space with Transformer-based long and short context modules, and a fast consistency MLP for frame prediction. Noise injection and Gaussian temperature scaling yield efficient and high-fidelity speech/music synthesis, outperforming discrete-token baselines both in quality (e.g., FAD vs. $1.06$/$1.43$ for MusicGen baselines) and generation speed (Simon et al., 8 Sep 2025).
- VideoMAR: Implements video modeling via continuous tokens with temporal causal autoregression and spatial masked prediction. Next-frame diffusion loss and curriculum strategies enable efficient, high-quality video generation with spatial-temporal extrapolation capacity, supported by 3D rotary position encodings and progressive temperature scheduling during inference (Yu et al., 17 Jun 2025).
- TarFlowLM: Employs transformer-based normalizing flows for continuous latent sequences of text. Mixture-of-Gaussian coupling layers provide bi-directional context, patch-wise flexible generation, and hierarchical multi-pass editing; theoretical analysis shows discrete AR transformers are a special case of the flow-based regime (Zhang et al., 1 Jul 2025).
5. Theoretical Foundations: Information Geometry, Spectral Analysis, and Surplus
Recent theoretical frameworks have illuminated why autoregressive models—including CALM—learn highly structured embedding spaces and enable multi-token acceleration.
- Markov Categorical Framework: AR generation is modeled compositionally as Markov kernels in ; NLL is equated to average KL divergence over kernels. Speculative decoding methods exploit the mutual information surplus in hidden states: justifying multi-token draft-and-verify acceleration (Zhang, 25 Jul 2025).
- Categorical Entropy: Minimizing NLL compels the model to reproduce the data’s intrinsic stochasticity ( convergence), not just deterministic predictions.
- Spectral Contrastive Learning: NLL implicitly aligns learned representations with the eigenspectrum of the predictive similarity operator, building geometrically and spectrally structured latent spaces. The Fisher-Rao pullback metric quantifies predictive sensitivity, and the Dirichlet energy of representation organization matches spectral objectives common to self-supervised contrastive learning.
- Implicit Continuity: Transformers trained on discrete data naturally learn representations giving rise to continuous-time and continuous-space invariance, with causal attention kernels generalizing from weighted sums to integrals over variable-duration, variable-semantic embedding sequences (Marro et al., 4 Apr 2025).
6. Empirical Performance, Scaling Laws, and Efficiency Implications
CALM demonstrates superior performance-compute trade-offs:
- CALM-M (371M params, ) matches Transformer-S (281M params) in BrierLM, with 44% fewer training FLOPs and 34% fewer inference FLOPs (Shao et al., 31 Oct 2025).
- In video, VideoMAR delivers SOTA fidelity with only 9.3% of the parameters, 0.5% of the training data, and 0.2% of the GPU resources compared to Cosmos I2V (Yu et al., 17 Jun 2025).
- Audio models achieve higher objective and subjective quality at generation speed (Simon et al., 8 Sep 2025).
A plausible implication is that increasing semantic bandwidth (chunk size or latent dimension ) provides a new axis for scaling laws, independent of model and data size—potentially unlocking further efficiency and capacity gains.
7. Design Principles, Future Directions, and Multimodal Extension
CALM principles yield several guidelines:
- Adopt high-fidelity, robust autoencoding as foundational.
- Model language, audio, and video as sequences of continuous latent vectors with autoregressive/flow-based priors.
- Employ strictly proper, likelihood-free scoring rules (energy/Brier) for both training and evaluation.
- Integrate flexible, likelihood-free sampling schemes, ensuring control over diversity/accuracy trade-offs.
- Exploit curriculum learning and progressive inference strategies for long/sequential modalities.
Potential future directions include scaling K-vector prediction to higher values with larger models, designing more expressive autoencoders/decoders, and extending CALM architectures to further modalities (e.g., image, code). Unified multimodal generative models based on continuous autoregression are anticipated, leveraging shared architectural and theoretical insights.
Key Mathematical Formulations in CALM
| Concept | Formula / Mechanism |
|---|---|
| Autoencoder Objective | |
| Variational KL | |
| AR Next-Vector | |
| Energy Score Loss | |
| BrierLM Metric |
Continuous Autoregressive LLMs represent a scalable, high-bandwidth, and theoretically rich approach to generative modeling. By shifting language modeling into a continuous latent domain, leveraging robust autoencoding, and employing likelihood-free statistical tools, CALM unlocks new frontiers for LLM scaling, multimodal synthesis, and efficient deployment.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free