Papers
Topics
Authors
Recent
Search
2000 character limit reached

Debiased LASSO under Poisson-Gauss Model

Published 26 Feb 2024 in math.ST, cs.IT, eess.SP, math.IT, and stat.TH | (2402.16764v1)

Abstract: Quantifying uncertainty in high-dimensional sparse linear regression is a fundamental task in statistics that arises in various applications. One of the most successful methods for quantifying uncertainty is the debiased LASSO, which has a solid theoretical foundation but is restricted to settings where the noise is purely additive. Motivated by real-world applications, we study the so-called Poisson inverse problem with additive Gaussian noise and propose a debiased LASSO algorithm that only requires $n \gg s\log2p$ samples, which is optimal up to a logarithmic factor.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. Variance-stabilization-based compressive inversion under poisson or poisson–gaussian noise with analytical bounds. Inverse Problems, 35(10):105006, 2019.
  2. Peter Bühlmann and Sara Van De Geer. Statistics for high-dimensional data: methods, theory and applications. Springer Science & Business Media, 2011.
  3. A method for modeling noise in medical images. IEEE Transactions on medical imaging, 23(10):1221–1232, 2004.
  4. High-dimensional confidence regions in sparse mri. In ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5, 2023a. doi: 10.1109/ICASSP49357.2023.10096320.
  5. Uncertainty quantification for learned ista. In 2023 IEEE 33rd International Workshop on Machine Learning for Signal Processing (MLSP), pages 1–6. IEEE, 2023b.
  6. A data-dependent weighted lasso under poisson noise. IEEE Transactions on Information Theory, 65(3):1589–1613, 2018.
  7. Confidence intervals and hypothesis testing for high-dimensional regression. The Journal of Machine Learning Research, 15(1):2869–2909, 2014.
  8. Debiasing the lasso: Optimal sample size for gaussian designs. The Annals of Statistics, 46(6A):2593–2622, 2018.
  9. Asymptotics in statistics: some basic concepts. Springer Science & Business Media, 2000.
  10. Sparsity regularization for image reconstruction with poisson data. In Computational Imaging VII, volume 7246, pages 96–105. SPIE, 2009.
  11. Gábor Lugosi. Lectures on combinatorial statistics. 47th Probability Summer School, Saint-Flour, pages 1–91, 2017.
  12. Sparse signal recovery under poisson statistics. In 2013 51st Annual Allerton Conference on Communication, Control, and Computing (Allerton), pages 1450–1457. IEEE, 2013.
  13. Compressed sensing performance bounds under poisson noise. IEEE Transactions on Signal Processing, 58(8):3990–4002, 2010.
  14. Taly Gilat Schmidt. Optimal “image-based” weighting for energy-resolved CT. Medical physics, 36(7):3018–3027, 2009.
  15. Astronomical data analysis and sparsity: From wavelets to compressed sensing. Proceedings of the IEEE, 98(6):1021–1030, 2009.
  16. Scaled sparse linear regression. Biometrika, 99(4):879–898, 2012.
  17. On asymptotically optimal confidence regions and tests for high-dimensional models. The Annals of Statistics, 42(3):1166–1202, 2014.
  18. Roman Vershynin. High-dimensional probability: An introduction with applications in data science, volume 47. Cambridge university press, 2018.
  19. Martin J Wainwright. High-dimensional statistics: A non-asymptotic viewpoint, volume 48. Cambridge university press, 2019.
  20. Platelets: a multiscale approach for recovering edges and surfaces in photon-limited medical imaging. IEEE Transactions on Medical Imaging, 22(3):332–350, 2003.
  21. Socioscope: Spatio-temporal signal recovery from social media. In Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2012, Bristol, UK, September 24-28, 2012. Proceedings, Part II 23, pages 644–659. Springer, 2012.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.