Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 78 tok/s
Gemini 2.5 Pro 56 tok/s Pro
GPT-5 Medium 34 tok/s Pro
GPT-5 High 33 tok/s Pro
GPT-4o 104 tok/s Pro
Kimi K2 187 tok/s Pro
GPT OSS 120B 451 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

Regularization by Denoising (RED)

Updated 6 October 2025
  • Regularization by Denoising (RED) is a framework that reformulates inverse imaging problems by explicitly embedding state-of-the-art denoisers into the optimization cost.
  • It offers a clear energy functional which integrates denoising algorithms directly, yielding theoretical guarantees of convexity and convergence under certain conditions.
  • RED demonstrates superior reconstruction performance in tasks like deblurring and super-resolution, and supports flexible optimization strategies such as gradient descent and ADMM.

Regularization by Denoising (RED) is a class of methodologies that reformulate inverse problems—particularly in imaging—by using modern denoising algorithms as the foundation for the regularizer in a unified optimization framework. RED distinguishes itself from traditional variational regularization by embedding state-of-the-art denoisers directly into the definition of the prior, promoting reconstructions that reside near the “fixed-point set” of the denoiser. The RED approach thereby shifts the role of the denoiser from a mere post-processing operator to a central component of the optimization cost. This has enabled significant advances in inverse imaging, owing to the strong empirical performance and adaptability of learned and model-based denoisers.

1. Core Methodology and Mathematical Framework

The principal idea in RED is to encode the prior via the denoising engine itself. Let f()f(\cdot) be a denoiser. RED defines the regularization penalty as

ρ(x)=12x(xf(x))\rho(x) = \frac{1}{2} x^\top (x - f(x))

where xRnx \in \mathbb{R}^n is the image variable. Under mild conditions on ff—notably, local homogeneity and strong passivity—this functional is convex and Fréchet differentiable, yielding the gradient

ρ(x)=xf(x)\nabla \rho(x) = x - f(x)

as the regularization force. The full objective for a generic linear inverse problem y=Hx+ηy = Hx + \eta (linear operator HH, measurements yy, noise η\eta) is formulated as

E(x)=(y,x)+λρ(x)=(y,x)+λ2x(xf(x))E(x) = \ell(y, x) + \lambda \rho(x) = \ell(y, x) + \frac{\lambda}{2} x^\top (x - f(x))

where (y,x)\ell(y, x) is the data-fidelity term, typically quadratic for AWGN (e.g., (1/2σ2)Hxy22(1/2\sigma^2)\|Hx-y\|_2^2), and λ\lambda is the regularization parameter. The stationary points are characterized by solving

(y,x)+λ(xf(x))=0\nabla \ell(y, x) + \lambda(x - f(x)) = 0

The explicit formulation of both the objective and its gradient enables seamless integration with general-purpose optimization algorithms.

2. Advantages over Implicit Plug-and-Play Schemes

Unlike the Plug-and-Play Priors (PnP) approaches—which use a denoiser “plugged into” an algorithm such as ADMM without an explicit underlying cost—RED constructs an explicit regularization functional. The prior is therefore not implicit but is wholly defined in closed form, greatly improving the transparency and theoretical justification of the resulting Bayesian or variational model. This explicitness facilitates the analysis of convexity, uniqueness, and convergence properties (under suitable conditions), and opens access to acceleration techniques or alternative optimizers beyond ADMM (Romano et al., 2016).

3. Optimization Strategies

A broad range of iterative schemes are enabled by the explicit RED objective:

  • Gradient Descent: Immediate application using E(x)\nabla E(x).
  • Conjugate Gradient / Subspace Methods: Exploitation of problem structure (e.g., circulant HH).
  • ADMM / Variable Splitting: The data-fidelity and prior terms can be split across variables, with the denoising step kept explicit in the update.
  • Fixed-Point Iteration: Solving (y,x)+λ(xf(x))=0\nabla \ell(y, x) + \lambda(x - f(x)) = 0 directly.
  • Weighted Proximal Methods (WPM): Generalization to variable-metric updates, leading to quasi-Newton acceleration of RED-type objectives (Hong et al., 2019).

The explicit form of the cost function and its gradient allows exploitation of any optimization procedure compatible with differentiable objectives, visibility into convergence rates, and the flexibility to accelerate inversion with vector extrapolation or adaptive step-size selection (Hong et al., 2018, Hong et al., 2019).

4. The Role and Requirements of the Denoiser

For the RED framework to retain its theoretical guarantees, the denoising operator must satisfy specific properties:

  • Local Homogeneity: f(cx)cf(x)f(cx) \approx c f(x) for cc near 1.
  • Strong Passivity (or Nonexpansivity): The spectral radius (or the Lipschitz constant) of the Jacobian JfJ_f is at most 1.
  • Symmetric Jacobian: For some theoretical results (gradient simplification, convexity proofs), Jf=JfJ_f = J_f^\top is required (Reehorst et al., 2018). In practice, high-performance denoisers such as TNRD, BM3D, or CNN-based architectures can be effective even if these conditions are only approximately satisfied. When these properties do not strictly hold, explicit cost-function interpretations may fail, but the fixed-point formulation and empirical effectiveness remain robust (Reehorst et al., 2018). Recent research has extended the framework to denoisers that are contractive or averaged operators, facilitating rigorous convergence even for complex, learned denoisers (Nair et al., 2022).

5. Applications and Empirical Performance

RED has demonstrated strong performance in a variety of inverse problems in imaging, including:

  • Image Deblurring: State-of-the-art results on uniform and Gaussian blur kernels with additive noise. RED restores sharpness and reduces noise beyond competing schemes such as NCSR or IDD-BM3D. Results show improved PSNR and favorable visual quality (Romano et al., 2016).
  • Super-Resolution: RED achieves reconstructions of high visual fidelity and quantitative SNR, especially when a strong denoiser is used in the regularizer.
  • Other Modalities: The explicit framework supports adaptation to compressed sensing, MRI, tomography, and non-imaging inverse problems, paving the way for applications across scientific imaging domains (Romano et al., 2016).
  • Comparison to PnP: RED often matches or outperforms PnP-ADMM approaches, with the additional benefit of a well-defined energy functional and more tractable parameter selection due to the clarity of the underlying objective.

6. Convergence Properties

Convexity and global optimality in RED are guaranteed when both the fidelity term and the denoiser-induced regularizer are convex—which is ensured by local homogeneity and strong passivity (or nonexpansivity) of the denoiser, and convexity of (y,x)\ell(y,x) (Romano et al., 2016). In these regimes, iterative minimization schemes converge to the unique global solution. The decoupling of prior and solution strategy increases overall robustness. By contrast, in PnP schemes with implicit priors, convergence may only be to a fixed point, with no guarantee of global minimization.

7. Research Directions and Extensions

Future research avenues highlighted in foundational RED papers include:

  • Generalization to broader classes of denoisers: Leveraging deep learning-based denoisers under controlled regularity conditions may yield further advances.
  • Adaptive and data-driven parameter selection: Dynamically adjusting denoiser parameters (e.g., noise level in f(x)f(x)) during optimization remains largely unexplored.
  • Higher-order and composite regularization: Penalizing higher moments of the residual xf(x)x - f(x) or introducing composite formulations for structured priors (e.g., group sparsity) represent promising directions.
  • Application to new problem domains: Tomographic reconstruction, segmentation, optical flow, and more general inverse problems may benefit from RED’s explicit cost and flexibility.
  • Handling non-differentiable or less regular denoisers: Theoretical advances continue to analyze RED under relaxed smoothness or symmetry conditions (Nair et al., 2022, Reehorst et al., 2018).
  • Parameter tuning methodologies: Empirical and theoretically motivated schemes for automatic λ\lambda selection, possibly using Stein’s Unbiased Risk Estimation (SURE), are of practical importance (Romano et al., 2016).

Table: Comparison of RED and Plug-and-Play (P³) Frameworks

Aspect RED PnP (P³)
Regularizer Explicit, cost function in closed form Implicit, no explicit objective
Prior Incorporation Direct via energy term ρ(x)\rho(x) Indirect via denoising step (usually ADMM)
Optimization Any gradient-based method Primarily ADMM, others less interpretable
Theoretical Props Convexity, convergence to global optimum (some) Fixed-point convergence; global optimality?
Parameter selection More transparent (via λ\lambda in objective) More heuristic, less clear

References

The RED framework provides a mathematically principled and engineering-flexible pathway for integrating learned and analytical denoisers into the solution of inverse problems. By leveraging the explicit structure of the regularizer defined through a denoising operator, RED enables convergence guarantees, adaptability across modalities, and the potential for further advances as new denoisers and optimization strategies are developed.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Regularization by Denoising (RED).

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube