Papers
Topics
Authors
Recent
2000 character limit reached

Beyond the Noise: Aligning Prompts with Latent Representations in Diffusion Models (2512.08505v1)

Published 9 Dec 2025 in cs.CV

Abstract: Conditional diffusion models rely on language-to-image alignment methods to steer the generation towards semantically accurate outputs. Despite the success of this architecture, misalignment and hallucinations remain common issues and require automatic misalignment detection tools to improve quality, for example by applying them in a Best-of-N (BoN) post-generation setting. Unfortunately, measuring the alignment after the generation is an expensive step since we need to wait for the overall generation to finish to determine prompt adherence. In contrast, this work hypothesizes that text/image misalignments can be detected early in the denoising process, enabling real-time alignment assessment without waiting for the complete generation. In particular, we propose NoisyCLIP a method that measures semantic alignment in the noisy latent space. This work is the first to explore and benchmark prompt-to-latent misalignment detection during image generation using dual encoders in the reverse diffusion process. We evaluate NoisyCLIP qualitatively and quantitatively and find it reduces computational cost by 50% while achieving 98% of CLIP alignment performance in BoN settings. This approach enables real-time alignment assessment during generation, reducing costs without sacrificing semantic fidelity.

Summary

  • The paper presents NoisyCLIP, a novel method for measuring text-to-image alignment in the latent space that halves computational cost while retaining 98% of alignment performance.
  • The method uses a twin-tower encoder architecture with contrastive fine-tuning on noisy latent representations to enhance semantic alignment during reverse diffusion.
  • Benchmark analysis shows improved prompt adherence and robust recall, setting a new standard for real-time, in-process alignment verification in generative image synthesis.

Language-to-Latent Alignment in Diffusion Models: A Formal Analysis of NoisyCLIP

Introduction and Motivation

Conditional diffusion models have become the backbone of contemporary image generation pipelines, delivering high-fidelity samples steered by textual input. Despite advances in cross-attention dynamics and latent-space manipulation, misalignment and semantic drift between prompts and outputs remain pervasive, particularly in computationally intensive scenarios like Best-of-N (BoN) selection. Existing evaluation protocols focus predominantly on post-generation alignment metrics, which incur substantial compute overhead due to the necessity of rendering all outputs before filtering for semantic fidelity. "Beyond the Noise: Aligning Prompts with Latent Representations in Diffusion Models" (2512.08505) addresses this bottleneck by introducing NoisyCLIP, a method for measuring text-to-image alignment in the latent space during diffusion, enabling real-time selection and early termination of poorly aligned samples. Figure 1

Figure 1: NoisyCLIP measures language-to-latent alignment during reverse diffusion by comparing textual prompts to intermediate latent representations, enabling early detection of misalignments.

Framework: NoisyCLIP and Latent Alignment Assessment

NoisyCLIP is built on a twin-tower encoder architecture, designed to operate within the noise-corrupted latent domains produced at intermediate steps of the reverse diffusion trajectory. This approach is grounded in augmenting CLIP-style models, which traditionally operate in RGB image space, via contrastive fine-tuning on noisy latents extracted from the reverse diffusion process.

The authors formalize the alignment signal Slatent(zt,y)=τθ(y)⋅νθ(Φ(zt))S_\text{latent}(z_t, y) = \tau_\theta(y) \cdot \nu_\theta(\Phi(z_t)), where Φ\Phi is a linear mapping transforming the latent ztz_t to a form ingestible by the image encoder νθ\nu_\theta, and τθ\tau_\theta denotes the frozen text encoder shared with the underlying diffusion architecture. Contrastive learning is performed in latents corresponding to early and mid-reverse diffusion steps, leveraging a curated Noisy-Conceptual-Captions dataset and controlling for positive/negative pairs via explicit batch-level in-batch negatives. Figure 2

Figure 2: Similarity between prompt and generated image is computed at intermediate denoising stages, with scores tightly correlated to scores obtained on the final image due to encoder fine-tuning on noisy latents.

Benchmarking and Methodological Contributions

The paper introduces two benchmarks for systematic evaluation: Noisy-Conceptual-Captions (caption selection and factual consistency) and Noisy-GenAI-Bench (BoN selection covering attribute, scene, spatial, and logical prompt types). These datasets catalogue not only final images but full latent trajectories, enabling granular analysis of prompt-image adherence across incremental denoising steps.

Key experimental tasks undertaken include:

  • Language-to-Latent Alignment: Ranking of images for a given prompt at each latent stage, quantifying separability using VQAScore.
  • Best-of-N Selection: Early selection of the most aligned latent path, assessed at varying diffusion costs.
  • Factual Consistency: Recall@1 evaluation requiring selection of correct prompts versus adversarially corrupted distractors.

Strong Results: Efficiency–Fidelity Trade-off in BoN

Empirical results underscore NoisyCLIP's computational benefits: by operating selection at the midpoint (latent ≈\approx20–30) rather than image-space finalization, compute is halved while retaining 98% of CLIP-based alignment performance (VQAScore). Figure 3 demonstrates that NoisyCLIP separates top-ranked candidates as early as latent 20, achieving a ∼\sim3% score gap. CLIP, absent noisy latent training, fails to maintain separation except at low-noise endpoints. Figure 3

Figure 3: Alignment measurement analysis throughout reverse diffusion reveals that NoisyCLIP distinguishes aligned from misaligned images by latent 20.

Figure 4

Figure 4

Figure 4

Figure 4

Figure 4: Alignment (VQAScore) plotted against denoising cost, showing NoisyCLIP achieves comparable alignment at half the computational cost (N=6 BoN).

Qualitative and quantitative results indicate superior early selection in BoN across both basic and advanced prompt categories, with notable improvements over baseline CLIP in complex, multi-entity or logically involved generations. Figure 5

Figure 5: Best-of-6 generation—NoisyCLIP outperforms CLIP at fixed budgets and often matches full-cost CLIP performance using half the diffusion steps.

Factual Consistency and Robustness

NoisyCLIP exhibits robust semantic discrimination, achieving Recall@1 scores exceeding 0.5 within the first half of denoising steps, outperforming CLIP across all tested non-factual distractor types (color, count, background, subject). Early correct caption selection is demonstrated in Figure 6. Figure 6

Figure 6: Recall@1 for correct versus non-factual distractor captions, indicating high discriminative power of NoisyCLIP at early latent stages.

Analysis across training latent intervals reveals optimal generalization for models fine-tuned in mid-range latent domains (20–29), balancing initial-stage robustness with terminal alignment accuracy, unlike low or high-noise exclusive training. Figure 7

Figure 7: Training/inference latent range correlation—moderate noise training yields stable alignment across the generation.

Ablation and Encoder Scalability

Experiments demonstrate that fine-tuning solely the image encoder suffices; joint text-image encoder tuning yields negligible or adverse gains. Encoder scaling studies show performance improvements proportional to model capacity (ViT-L/14 > ViT-B/16), solidifying the link between representational expressivity and robust alignment under noise.

Practical and Theoretical Implications

Practically, NoisyCLIP redefines BoN strategies in generative frameworks, offering early misalignment detection, dynamic early stopping, and compute savings whilst maintaining semantic quality. This paradigm enables integration of robust alignment metrics into control loops for real-time, efficient multimodal generation. Theoretically, the study evidences that latent-space alignment metrics, when fine-tuned on noise-corrupted trajectories, can reliably project final semantic fidelity earlier in the generative process. Extension to alternative contrastive backbones (e.g. SigLIP) shows promise but highlights architectural variance in transferability.

Conclusion

NoisyCLIP operationalizes continuous language-to-latent alignment evaluation throughout reverse diffusion, transforming semantic validation from a post-hoc filter into a computationally integral selection mechanism. With strong evidence for significant cost reductions and minimal fidelity loss, this approach establishes a new baseline for in-process alignment-aware generation and informs future directions for adaptive prompt-conditioned control in scalable diffusion pipelines. The open benchmarks and protocol blueprints provided invite further research in latent diagnostics, backbone generalization, and transfer to broader multimodal frameworks.

Whiteboard

Video Overview

Open Problems

We found no open problems mentioned in this paper.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 18 likes about this paper.