Papers
Topics
Authors
Recent
2000 character limit reached

Generalized Denoising Diffusion Compression

Updated 24 November 2025
  • gDDCM is a unified framework for image compression and tokenization using diffusion-based generative models and fixed Gaussian codebooks.
  • It replaces continuous noise injection with quantization from a pre-sampled Gaussian codebook, enabling lossless bit-stream representation and efficient discrete tokenization.
  • The model supports tasks like restoration and super-resolution, achieving state-of-the-art perceptual metrics on benchmarks such as CIFAR-10 and LSUN Bedroom.

The Generalized Denoising Diffusion Compression Model (gDDCM) is a unified framework for image compression and tokenization grounded in diffusion generative models. It extends the Denoising Diffusion Codebook Model (DDCM), originally designed for discrete-time denoising diffusion probabilistic models (DDPM), to encompass a broad family of diffusion model variants, including score-based generative models, consistency models, and rectified flow. The core innovation is replacing the continuous stochastic noise injection at each reverse diffusion step with quantization over a fixed, small Gaussian codebook, enabling lossless bit-stream representation and a discrete token sequence per image. This mechanism operates at inference time without retraining the underlying generative model and supports both unconditional compression and conditional image generation tasks such as restoration and super-resolution, with state-of-the-art perceptual metrics across datasets (Ohayon et al., 3 Feb 2025, Kong, 17 Nov 2025).

1. Theoretical Foundation and Model Architecture

gDDCM is built upon diffusion-based generative models where the marginal at step tt can be expressed as xt=s(t)x0+σ(t)ϵx_t = s(t) x_0 + \sigma(t) \epsilon, with ϵN(0,I)\epsilon \sim \mathcal{N}(0, I) and s(t),σ(t)s(t), \sigma(t) denoting the known scalar schedules specific to the chosen diffusion paradigm (Kong, 17 Nov 2025). The forward process remains standard (e.g., for DDPM: xt=1βtxt1+βtϵtx_t = \sqrt{1-\beta_t} x_{t-1} + \sqrt{\beta_t} \epsilon_t).

The key architectural components are:

  • Pre-trained backbone: gDDCM utilizes any diffusion-family network (DDPM/DDIM, score-based SDE, consistency model, rectified flow) without modification or retraining, relying on the original score estimation (e.g., ϵθ(xt,t)\epsilon_\theta(x_t,t) or sθ(xt,t)s_\theta(x_t,t)) learned under standard objectives (Ohayon et al., 3 Feb 2025, Kong, 17 Nov 2025).
  • Fixed Gaussian codebooks: At each of NN discrete or continuous time-steps, a codebook St={e(1),,e(M)}\mathcal{S}_t = \{e^{(1)}, \ldots, e^{(M)}\} of i.i.d. Gaussian vectors is pre-sampled and fixed. These serve as the finite set from which noise is selected, replacing fresh N(0,I)\mathcal{N}(0, I) draws at each reverse step (Ohayon et al., 3 Feb 2025, Kong, 17 Nov 2025).

2. Modified Sampling and Compression Mechanism

In classical DDPM reverse dynamics, a fresh noise draw is injected at each iteration:

xt1=μθ(xt,t)+Σθ(xt,t)ϵ,  ϵN(0,I)x_{t-1} = \mu_\theta(x_t,t) + \Sigma_\theta(x_t,t)\epsilon,\; \epsilon \sim \mathcal{N}(0, I)

gDDCM amends this by substituting ϵ\epsilon with the nearest codebook entry zz:

xt1=μθ(xt,t)+Σθ(xt,t)z,z=argmineStϵe2x_{t-1} = \mu_\theta(x_t, t) + \Sigma_\theta(x_t, t) z,\quad z = \operatorname*{argmin}_{e\in \mathcal{S}_t} \|\epsilon' - e\|_2

where ϵ\epsilon' is the estimated noise at step tt (e.g., (xts(t)x0)/σ(t)(x_t - s(t)x_0)/\sigma(t) or from a model prediction) (Kong, 17 Nov 2025). The index ctc_t of zz is stored, yielding a discrete token sequence.

For continuous-time or flow-based variants (score-based SDE, consistency, rectified flow), a de-noising and re-noising strategy integrates codebook quantization into the ODE or SDE update, maintaining the same discretization convention via the unified marginal form (Kong, 17 Nov 2025).

3. Bit-Stream Construction and Rate–Distortion Analysis

The compressed representation is the sequence of codebook indices {i1,...,iN}\{i_1, ..., i_N\} per image. Each index requires log2M\log_2 M bits; total bitrate is Nlog2MN \log_2 M bits per image. Bitrate per dimension (bpd\mathrm{bpd}) is calculated as:

bpd=Nlog2MHWC\mathrm{bpd} = \frac{N \log_2 M}{H \cdot W \cdot C}

where H,W,CH, W, C are height, width, and channel count. Entropy coding (e.g., Huffman, arithmetic) can further reduce bit-cost by leveraging non-uniform token probabilities (Kong, 17 Nov 2025).

Rate–distortion tradeoff is tunable via codebook size MM and step count NN:

  • Larger MM increases fidelity but also token entropy.
  • Gains saturate beyond M512M \approx 512 and N300N \approx 300 on typical datasets such as CIFAR-10 (Kong, 17 Nov 2025).
  • Matching pursuit and convex codebook combinations permit finer bitrate control at the cost of additional encoding bits (Ohayon et al., 3 Feb 2025).

4. Generalization to Conditional Generation and Tasks

gDDCM generalizes tokenized inference to arbitrary diffusion-based conditional generation by integrating a task-specific loss (y,xi,z)\ell(y, x_i, z) into the codebook selection:

ki=argmin1kK(y,xi,zi(k))k_i = \operatorname*{argmin}_{1\leq k \leq K} \ell(y, x_i, z_i^{(k)})

With suitable \ell, gDDCM recovers various tasks:

  • Lossy compression: Set y=x0y = x_0 and \ell as a posterior-inspired quadratic (e.g., P=zσixilogpi(xiy)2\ell_P = \|z - \sigma_i \nabla_{x_i} \log p_i(x_i|y)\|^2).
  • Restoration/Super-resolution: Losses such as (xi+σiz)g(y)2-\|(x_i + \sigma_i z) - g(y)\|^2 bias toward reconstructions consistent with degraded inputs (Ohayon et al., 3 Feb 2025). As KK \to \infty, the quantized noise converges to the posterior score, and the procedure discretizes the probability-flow ODE, yielding approximate posterior sampling (Ohayon et al., 3 Feb 2025).

5. Practical Implementation and Experimental Findings

gDDCM is entirely inference-driven: no retraining, fine-tuning, or auxiliary loss terms are needed. The only computational overhead is per-step nearest neighbor search in the codebook (Ohayon et al., 3 Feb 2025, Kong, 17 Nov 2025). The encoding/decoding protocol involves re-generating Gaussian codebooks from a shared seed, quantizing estimated noise vectors, and reconstructing the reverse process.

On standard benchmarks:

  • CIFAR-10 (32×32×3,N=300,M=25632\times32\times3, N=300, M=256): gDDCM with p=0p=0 achieves FID=3.2, LPIPS=0.060, IS=10.5, SSIM=0.98. Baseline DDCM with p=0.5p=0.5 yields FID=7.7, SSIM=0.93 (Kong, 17 Nov 2025).
  • LSUN Bedroom (256×256×3,N=600,M=512256\times256\times3, N=600, M=512): gDDCM attains LPIPS 0.052\approx0.052, SSIM 0.96\approx0.96 (Kong, 17 Nov 2025).
  • gDDCM achieves state-of-the-art performance at extreme low bitrates (\sim0.05 BPP), surpassing methods such as BPG, HiFiC, PSC, and PerCo in FID and LPIPS metrics, while rates at higher bitrates are capped by the underlying VAE latent space (Ohayon et al., 3 Feb 2025).

The approach demonstrates uniform applicability across DDPM, score-based SDE (EDM), consistency, and flow-matching models, consistently improving FID and SSIM relative to baseline DDCM.

6. Limitations, Implications, and Open Directions

The only overhead introduced by gDDCM is the codebook search per step. The quality–bitrate tradeoff and codebook efficiency are currently limited by the choice of MM and NN, with no further reduction in bits after saturation. A plausible implication is that per-step adaptive or jointly learned codebooks could further compress image representations. Future research directions explicitly identified include:

  • Joint learning of codebooks to minimize MM without loss of quality.
  • Incorporating entropy-regularized or variational token selection.
  • Extension to conditional or structured data generation (e.g., multi-modal, class-conditional tasks).
  • Theoretical rate–distortion analysis specific to diffusion-based codecs (Kong, 17 Nov 2025).
  • Optimizing quantization strategies that align with human perceptual metrics.

gDDCM provides a principled, model-agnostic tokenization methodology for generative compression and conditional generation across the spectrum of diffusion-based models, achieving efficient and high-fidelity discrete representations on challenging visual benchmarks (Ohayon et al., 3 Feb 2025, Kong, 17 Nov 2025).

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Generalized Denoising Diffusion Compression Model (gDDCM).