Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
166 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Listening to the Noise: Blind Denoising with Gibbs Diffusion (2402.19455v2)

Published 29 Feb 2024 in stat.ML, astro-ph.CO, cs.CV, cs.LG, and eess.SP

Abstract: In recent years, denoising problems have become intertwined with the development of deep generative models. In particular, diffusion models are trained like denoisers, and the distribution they model coincide with denoising priors in the Bayesian picture. However, denoising through diffusion-based posterior sampling requires the noise level and covariance to be known, preventing blind denoising. We overcome this limitation by introducing Gibbs Diffusion (GDiff), a general methodology addressing posterior sampling of both the signal and the noise parameters. Assuming arbitrary parametric Gaussian noise, we develop a Gibbs algorithm that alternates sampling steps from a conditional diffusion model trained to map the signal prior to the family of noise distributions, and a Monte Carlo sampler to infer the noise parameters. Our theoretical analysis highlights potential pitfalls, guides diagnostic usage, and quantifies errors in the Gibbs stationary distribution caused by the diffusion model. We showcase our method for 1) blind denoising of natural images involving colored noises with unknown amplitude and spectral index, and 2) a cosmology problem, namely the analysis of cosmic microwave background data, where Bayesian inference of "noise" parameters means constraining models of the evolution of the Universe.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (65)
  1. Stochastic interpolants: A unifying framework for flows and diffusions. arXiv preprint arXiv:2303.08797, 2023.
  2. The RWST, a comprehensive statistical description of the non-Gaussian structures in the ISM. Astronomy & Astrophysics, 629:A115, 2019. doi: 10.1051/0004-6361/201834975.
  3. Anderson, B. D. Reverse-time diffusion equation models. Stochastic Processes and their Applications, 12(3):313–326, 1982.
  4. Conditionally specified distributions: an introduction. Statistical Science, 16:249–274, 2001.
  5. Cleaning our own dust: simulating and separating galactic dust foregrounds with neural networks. Monthly Notices of the Royal Astronomical Society, 500(3):3889–3897, 2021. doi: 10.1093/mnras/staa3344.
  6. Noise2Self: Blind Denoising by Self-Supervision. In Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pp.  524–533. PMLR, 2019.
  7. Betancourt, M. Nested Sampling with Constrained Hamiltonian Monte Carlo. AIP Conference Proceedings, 1305(1):165–172, 03 2011. ISSN 0094-243X. doi: 10.1063/1.3573613.
  8. Betancourt, M. A Conceptual Introduction to Hamiltonian Monte Carlo. arXiv preprint arXiv:1701.02434, 2018.
  9. Joint Analysis of BICEP2/Keck Array and Planck Data. Physical Review Letters, 114(10):101301, 2015. doi: 10.1103/PhysRevLett.114.101301.
  10. The Catalogue for Astrophysical Turbulence Simulations (CATS). The Astrophysical Journal, 905(1):14, 2020.
  11. Monte Carlo guided Denoising Diffusion models for Bayesian linear inverse problems. In The Twelfth International Conference on Learning Representations, 2024.
  12. Diffusion Posterior Sampling for General Noisy Inverse Problems. In The Eleventh International Conference on Learning Representations, 2023.
  13. Image denoising by sparse 3D transform-domain collaborative filtering. IEEE Transactions on Image Processing, 16(8):2080 – 2095, 2007. ISSN 22195491.
  14. ImageNet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp.  248–255, 2009. doi: 10.1109/CVPR.2009.5206848.
  15. Efron, B. Tweedie’s Formula and Selection Bias. Journal of the American Statistical Association, 106(496):1602–1614, 2011. doi: 10.1198/jasa.2011.tm11181.
  16. Blind Universal Bayesian Image Denoising With Gaussian Noise Level Learning. IEEE Transactions on Image Processing, 29:4885–4897, 2020. doi: 10.1109/TIP.2020.2976814.
  17. Image Denoising: The Deep Learning Revolution and Beyond—A Survey Paper. SIAM Journal on Imaging Sciences, 16(3):1594–1654, 2023. doi: 10.1137/23M1545859.
  18. Stochastic Relaxation, Gibbs distributions, and the Bayesian restoration of images. IEEE Transactions on Pattern Analysis and Machine Intelligence, 6:721–741, 1984.
  19. On the characterization of a joint probability distribution by conditional distributions. Journal of Econometrics, 10:115–118, 1979.
  20. Removing Dust from CMB Observations with Diffusion Models. NeurIPS Workshop on Machine Learning and the Physical Sciences, 2023. doi: 10.48550/arXiv.2310.16285.
  21. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840–6851, 2020.
  22. Functional compatibility, Markov chains and Gibbs sampling with improper posteriors. Journal of Computational and Graphical Statistics, 7:42–60, 1998.
  23. Solving high-dimensional parameter inference: marginal posterior densities & Moment Networks. In NeurIPS Workshop on Machine Learning and the Physical Sciences, 2020.
  24. Solving Linear Inverse Problems Using the Prior Implicit in a Denoiser. arXiv preprint arXiv:2007.13640, 2021.
  25. The quest for b modes from inflationary gravitational waves. Annual Review of Astronomy and Astrophysics, 54(1):227–269, 2016. doi: 10.1146/annurev-astro-081915-023433.
  26. Denoising diffusion restoration models. In Advances in Neural Information Processing Systems, 2022.
  27. On Density Estimation with Diffusion Models. In Advances in Neural Information Processing Systems, 2021.
  28. ForSE: A GAN-based Algorithm for Extending CMB Foreground Models to Subdegree Angular Scales. The Astrophysical Journal, 911(1):42, 2021. doi: 10.3847/1538-4357/abe71c.
  29. Efficient Computation of Cosmic Microwave Background Anisotropies in Closed Friedmann-Robertson-Walker Models. The Astrophysical Journal, 538(2):473, 2000. doi: 10.1086/309179.
  30. On the Stationary Distribution of Iterative Imputations. Biometrika, 101:155–173, 2012.
  31. On the geometric ergodicity of Hamiltonian Monte Carlo. Bernoulli, 25:3109–3138, 2019.
  32. Decoupled Weight Decay Regularization. In International Conference on Learning Representations, 2019.
  33. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001, volume 2, pp.  416–423. IEEE, 2001.
  34. Diffusion Model Based Posterior Samplng for Noisy Linear Inverse Problems. arXiv preprint arXiv:2211.12343, 2022.
  35. Can denoising diffusion probabilistic models generate realistic astrophysical fields? arXiv preprint arXiv:2211.12444, 2022.
  36. GibbsDDRM: A partially collapsed Gibbs sampler for solving blind inverse problems with denoising diffusion restoration. In Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp.  25501–25522. PMLR, 2023.
  37. Collaborative Filtering of Correlated Noise: Exact Transform-Domain Variance for Improved Shrinkage and Patch Matching. IEEE Transactions on Image Processing, 29:8339–8354, 2020. doi: 10.1109/TIP.2020.3014721.
  38. Neal, R. M. MCMC using Hamiltonian dynamics. Handbook of Markov Chain Monte Carlo, 54:113–162, 2010.
  39. Nesterov, Y. Primal-dual subgradient methods for convex problems. Mathematical Programming, 120(1):221–259, 2009. doi: 10.1007/s10107-007-0149-x.
  40. Planck Collaboration XI. Planck 2013 results. XI. All-sky model of thermal dust emission. Astronomy & Astrophysics, 571:A11, 2014. doi: 10.1051/0004-6361/201323195.
  41. Planck Collaboration I. Planck 2018 results. I. Overview, and the cosmological legacy of Planck. Astronomy & Astrophysics, 641:A1, 2020. doi: 10.1051/0004-6361/201833880.
  42. Planck Collaboration IV. Planck 2018 results. IV. Diffuse component separation. Astronomy & Astrophysics, 641:A4, 2020. doi: 10.1051/0004-6361/201833881.
  43. Statistical description of dust polarized emission from the diffuse interstellar medium. A RWST approach. Astronomy & Astrophysics, 642:A217, 2020. doi: 10.1051/0004-6361/202038044.
  44. A new approach for the statistical denoising of Planck interstellar dust polarization data. Astronomy & Astrophysics, 649:L18, 2021. doi: 10.1051/0004-6361/202140503.
  45. Generative Models of Multichannel Data from a Single Example-Application to Dust Emission. The Astrophysical Journal, 943(1):9, 2023. doi: 10.3847/1538-4357/aca538.
  46. Robbins, H. E. An Empirical Bayes Approach to Statistics. 1956.
  47. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  10684–10695, 2022.
  48. Solving Linear Inverse Problems Provably via Posterior Sampling with Latent Diffusion Models. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  49. Imagenet large scale visual recognition challenge. International journal of computer vision, 115:211–252, 2015.
  50. Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding. In Advances in Neural Information Processing Systems, 2022.
  51. Deep Unsupervised Learning using Nonequilibrium Thermodynamics. In Proceedings of the 32nd International Conference on Machine Learning, volume 37 of Proceedings of Machine Learning Research, pp.  2256–2265. PMLR, 2015.
  52. Pseudoinverse-Guided Diffusion Models for Inverse Problems. In International Conference on Learning Representations, 2023a.
  53. Score-Based Generative Modeling through Stochastic Differential Equations. In International Conference on Learning Representations, 2021.
  54. Consistency models. In Proceedings of the 40th International Conference on Machine Learning, 2023b.
  55. Removing Structured Noise with Diffusion Models. arXiv preprint arXiv:2302.05290, 2023.
  56. Validating Bayesian Inference Algorithms with Simulation-Based Calibration. arXiv preprint arXiv:1804.06788, 2018.
  57. A generative model of galactic dust emission using variational autoencoders. Monthly Notices of the Royal Astronomical Society, 504(2):2603–2613, 2021. doi: 10.1093/mnras/stab1011.
  58. Deep learning on image denoising: An overview. Neural Networks, 131:251–275, 2020. ISSN 0893-6080. doi: https://doi.org/10.1016/j.neunet.2020.07.025.
  59. Statistical simulations of the dust foreground to cosmic microwave background polarization. Astronomy & Astrophysics, 603:A62, 2017. doi: 10.1051/0004-6361/201629992.
  60. Rank-Normalization, Folding, and Localization: An Improved R^^𝑅\widehat{R}over^ start_ARG italic_R end_ARG for Assessing Convergence of MCMC (with Discussion). Bayesian Analysis, 16(2):667 – 718, 2021. doi: 10.1214/20-BA1221.
  61. Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing, 13(4):600–612, 2004. doi: 10.1109/TIP.2003.819861.
  62. Diffusion Model for Generative Image Denoising. arXiv preprint arXiv:2302.02398, 2023.
  63. Beyond a Gaussian denoiser: Residual learning of deep CNN for image denoising. IEEE Transactions on Image Processing, 26(7):3142–3155, 2017.
  64. Plug-and-Play Image Restoration With Deep Denoiser Prior. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(10):6360–6376, 2022. doi: 10.1109/TPAMI.2021.3088914.
  65. Denoising Diffusion Models for Plug-and-Play Image Restoration. In IEEE Conference on Computer Vision and Pattern Recognition Workshops (NTIRE), 2023.
Citations (1)

Summary

  • The paper presents Gibbs Diffusion, integrating diffusion models with Gibbs sampling to address blind denoising challenges.
  • It efficiently estimates both signal and noise parameters through Monte Carlo sampling and diagonalizes covariance in Fourier space.
  • Empirical results on natural images and cosmic microwave background data reveal superior performance compared to established methods.

Overview of "Listening to the Noise: Blind Denoising with Gibbs Diffusion"

The paper "Listening to the Noise: Blind Denoising with Gibbs Diffusion" presents an innovative approach named Gibbs Diffusion (GDiff) to tackle the challenge of blind denoising, which involves inferring both the original signal and the characteristics of noise from observed data. This research intersects fields such as machine learning, physics, and image processing, offering a new perspective on applying deep generative models for denoising tasks.

Key Contributions

The primary contribution of the work is the introduction of GDiff, a combination of diffusion models and a Gibbs sampler. This method enhances the blind denoising process by concurrently sampling both signal and noise parameters. The methodology is constructed on the foundation of diffusion models which have been shown to effectively approximate complex data distributions and are already prominent in state-of-the-art generative models. Notably, GDiff manages the intricacies of noise covariance by incorporating a Monte Carlo sampling approach, allowing inference on noise parameters under a Bayesian framework. The flexibility of GDiff is demonstrated in two significant applications: denoising of natural images with colored noise and a cosmological analysis of cosmic microwave background (CMB) data.

Methodology

The methodology employs a diffusion model trained to reverse a stochastic process that progressively adds noise to data, effectively serving as a mechanism to sample from the posterior distribution over the denoised signal. This requires modeling noise variance and covariance, and the paper details a conditional model that accounts for unknown noise parameters. The practical implementation involves a Gibbs sampler alternately drawing from conditional distributions over signals and noise parameters. To handle high-dimensional noise covariance within a manageable parameter space, the researchers emphasize the diagonalization of the covariance matrix in Fourier space, which reduces computational complexity considerably.

Theoretical Insights

The authors provide a comprehensive theoretical analysis of their approach, addressing potential pitfalls and ensuring the robustness of the Gibbs Diffusion model. This includes conditions for the stationary distribution of the Gibbs sampler, as well as diagnostics for tuning the inference and quantifying the propagation of errors. Importantly, they highlight the existence of a reverse stochastic process which aids in the denoising and posterior sampling procedure, supported by the theoretical backing from prior work on stochastic differential equations.

Numerical Results and Applications

The authors demonstrate the effectiveness of their method with empirical results. For natural images contaminated with noises of unknown amplitude, GDiff, despite operating blindly, achieves superior denoising performance compared to established methods like BM3D and DnCNN in certain metrics such as peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM). Additionally, the cosmological application showcases the method's utility in decomposing CMB data into foreground and background components, a complex task due to intricate interferences in cosmological observations. These results underscore the potential of GDiff to enhance scientific inferences in cosmology by accurately modeling the signal and noise interplay.

Implications and Speculations

Practically, GDiff expands the toolkit available for addressing noisy signal reconstruction problems, particularly in domains where noise characteristics are inherently unknown or variable. Theoretically, it situates itself as a bridge integrating diffusion processes and Bayesian inference, which may inspire further explorations into more versatile generative models capable of posterior sampling beyond Gaussian assumptions. The versatility demonstrated by addressing both image processing and cosmological data highlights its applicability across a breadth of scientific disciplines.

From a speculative viewpoint, future works may build upon this foundation by integrating more sophisticated adaptation schemes for Gibbs sampling or exploring extensions that relax Gaussian noise assumptions. Additionally, refining the computational efficiency of diffusion models, particularly for real-time applications, remains an open challenge.

In summary, "Listening to the Noise: Blind Denoising with Gibbs Diffusion" introduces a robust, flexible framework balancing state-of-the-art techniques in generative modeling with the demands of nuanced scientific data analysis, paving the way for more dynamic and adaptive noise processing strategies.

Youtube Logo Streamline Icon: https://streamlinehq.com