Quantum Algorithms for the Pathwise Lasso (2312.14141v2)
Abstract: We present a novel quantum high-dimensional linear regression algorithm with an $\ell_1$-penalty based on the classical LARS (Least Angle Regression) pathwise algorithm. Similarly to available classical algorithms for Lasso, our quantum algorithm provides the full regularisation path as the penalty term varies, but quadratically faster per iteration under specific conditions. A quadratic speedup on the number of features $d$ is possible by using the quantum minimum-finding routine from D\"urr and Hoyer (arXiv'96) in order to obtain the joining time at each iteration. We then improve upon this simple quantum algorithm and obtain a quadratic speedup both in the number of features $d$ and the number of observations $n$ by using the approximate quantum minimum-finding routine from Chen and de Wolf (ICALP'23). As one of our main contributions, we construct a quantum unitary to approximately compute the joining times to be searched over by the approximate quantum minimum finding. Since the joining times are no longer exactly computed, it is no longer clear that the resulting approximate quantum algorithm obtains a good solution. As our second main contribution, we prove, via an approximate version of the KKT conditions and a duality gap, that the LARS algorithm (and thus our quantum algorithm) is robust to errors. This means that it still outputs a path that minimises the Lasso cost function up to a small error if the joining times are approximately computed. Moreover, we show that, when the observations are sampled from a Gaussian distribution, our quantum algorithm's complexity only depends polylogarithmically on $n$, exponentially better than the classical LARS algorithm, while keeping the quadratic improvement on $d$. Finally, we propose a dequantised algorithm that also retains the polylogarithmic dependence on $n$, albeit with the linear scaling on $d$ from the standard LARS algorithm.
- Constant-depth circuits for Uniformly Controlled Gates and Boolean functions with application to quantum memory circuits. arXiv preprint arXiv:2308.08539, 2023.
- Efficient implementations of the generalized lasso dual path algorithm. Journal of Computational and Graphical Statistics, 25(1):1–27, 2016.
- Estimating lasso risk and noise level. Advances in Neural Information Processing Systems, 26, 2013.
- Quantum amplitude amplification and estimation. Contemporary Mathematics, 305:53–74, 2002.
- Convex Analysis. Springer, 2006.
- Forward-backward selection with early dropping. The Journal of Machine Learning Research, 20(1):276–314, 2019.
- Peter Bühlmann and Sara van de Geer. Statistics for high-dimensional data: methods, theory and applications. Springer Science & Business Media, 2011.
- Quantum matching pursuit: A quantum algorithm for sparse representations. Physical Review A, 105(2):022414, 2022.
- Condition numbers of Gaussian random matrices. SIAM Journal on Matrix Analysis and Applications, 27(3):603–620, 2005.
- Atomic decomposition by basis pursuit. SIAM review, 43(1):129–159, 2001.
- Quantum algorithms and lower bounds for linear regression with norm constraints. In 50th International Colloquium on Automata, Languages, and Programming (ICALP 2023), volume 261 of Leibniz International Proceedings in Informatics (LIPIcs), pages 38:1–38:21, Dagstuhl, Germany, 2023. Schloss Dagstuhl – Leibniz-Zentrum für Informatik.
- The power of block-encoded matrix powers: Improved regression techniques via faster hamiltonian simulation. Proceedings of the 46th International Colloquium on Automata, Languages, and Programming (ICALP), 132, 2019.
- Quantum regularized least squares. Quantum, 7:988, 2023.
- Near-ideal model selection by ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT minimization. The Annals of Statistics, 37(5A):2145–2177, 2009.
- Stable signal recovery from incomplete and inaccurate measurements. Communications on Pure and Applied Mathematics: A Journal Issued by the Courant Institute of Mathematical Sciences, 59(8):1207–1223, 2006.
- Faster quantum ridge regression algorithm for prediction. International Journal of Machine Learning and Cybernetics, 14(1):117–124, 2023.
- A quantum algorithm for finding the minimum. arXiv preprint quant-ph/9607014, 1996.
- Uncertainty principles and ideal atomic decomposition. IEEE transactions on information theory, 47(7):2845–2862, 2001.
- Algorithms for subset selection in linear regression. In Proceedings of the fortieth annual ACM symposium on Theory of computing, pages 45–54, 2008.
- David L. Donoho. For most large underdetermined systems of linear equations the minimal ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT-norm solution is also the sparsest solution. Communications on Pure and Applied Mathematics: A Journal Issued by the Courant Institute of Mathematical Sciences, 59(6):797–829, 2006.
- Ashok Vithoba Dorugade. New ridge parameters for ridge regression. Journal of the Association of Arab Universities for Basic and Applied Sciences, 15:94–99, 2014.
- Charles Dossal. A necessary and sufficient condition for exact sparse recovery by ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT minimization. Comptes Rendus Mathematique, 350(1-2):117–120, 2012.
- A generalized uncertainty principle and sparse representation in pairs of bases. IEEE Transactions on Information Theory, 48(9):2558–2567, 2002.
- Least angle regression. The Annals of Statistics, 32(2):407 – 499, 2004.
- Tails of condition number distributions. SIAM journal on matrix analysis and applications, 27(2):547–560, 2005.
- Regularization paths for generalized linear models via coordinate descent. Journal of statistical software, 33(1):1, 2010.
- A note on the group lasso and a sparse group lasso. arXiv preprint arXiv:1001.0736, 2010.
- A selective overview of variable selection in high dimensional feature space. Statistica Sinica, 20(1):101, 2010.
- Gradient projection for sparse reconstruction: Application to compressed sensing and other inverse problems. IEEE Journal of selected topics in signal processing, 1(4):586–597, 2007.
- An invitation to compressive sensing. Springer, 2013.
- J.-J. Fuchs. Recovery of exact sparse representations in the presence of noise. In 2004 IEEE International Conference on Acoustics, Speech, and Signal Processing, volume 2, pages ii–533. IEEE, 2004.
- J.J. Fuchs. Recovery of exact sparse representations in the presence of bounded noise. IEEE Transactions on Information Theory, 51(10):3601–3608, 2005.
- An homotopy algorithm for the Lasso with online observations. Advances in neural information processing systems, 21, 2008.
- Algorithms for fitting the constrained lasso. Journal of Computational and Graphical Statistics, 27(4):861–871, 2018.
- Architectures for a quantum random access memory. Physical Review A, 78(5):052310, 2008.
- Quantum random access memory. Physical review letters, 100(16):160501, 2008.
- Quantum singular value transformation and beyond: exponential improvements for quantum matrix arithmetics. In Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing, pages 193–204, 2019.
- Steve R. Gunn. Support vector machines for classification and regression. Technical Report 1, University of Southampton, 1998.
- Quantum algorithm for linear systems of equations. Physical review letters, 103(15):150502, 2009.
- Ridge regression: Biased estimation for nonorthogonal problems. Technometrics, 12(1):55–67, 1970.
- Ridge regression: some simulations. Communications in Statistics-Theory and Methods, 4(2):105–123, 1975.
- Adaptive lasso for sparse high-dimensional regression models. Statistica Sinica, pages 1603–1618, 2008.
- Holger Hoefling. A path algorithm for the fused lasso signal approximator. Journal of Computational and Graphical Statistics, 19(4):984–1006, 2010.
- Statistical Learning with Sparsity: The Lasso and Generalizations. Chapman and Hall/CRC, May 2015.
- Statistical challenges of high-dimensional data, 2009.
- B.M. Golam Kibria. Performance of some new ridge regression estimators. Communications in Statistics-Simulation and Computation, 32(2):419–435, 2003.
- A gradient-based optimization algorithm for LASSO. Journal of Computational and Graphical Statistics, 17(4):994–1009, 2008.
- Lower bounds on randomly preconditioned Lasso via robust sparse designs. Advances in Neural Information Processing Systems, 35:24419–24431, 2022.
- Feature selection: a literature review. SmartCR, 4(3):211–229, 2014.
- Linear regression by quantum amplitude estimation and its extension to convex optimization. Physical Review A, 104(2):022430, 2021.
- Quantum recommendation systems. In 8th Innovations in Theoretical Computer Science Conference (ITCS 2017), volume 67 of Leibniz International Proceedings in Informatics (LIPIcs), pages 49:1–49:21, Dagstuhl, Germany, 2017. Schloss Dagstuhl – Leibniz-Zentrum für Informatik.
- Quantum gradient descent for linear systems and least squares. Physical Review A, 101(2):022316, 2020.
- Hamiltonian simulation by qubitization. Quantum, 3:163, 2019.
- Quantum principal component analysis. Nature Physics, 10(9):631–633, 2014.
- A significance test for the lasso. The Annals of Statistics, 42(2):413, 2014.
- K.Z. Mao. Fast orthogonal forward selection algorithm for feature subset selection. IEEE Transactions on Neural Networks, 13(5):1218–1224, 2002.
- K.Z. Mao. Orthogonal forward selection and backward elimination algorithms for feature subset selection. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 34(1):629–634, 2004.
- High-dimensional graphs and variable selection with the Lasso. The Annals of Statistics, 34(3):1436 – 1462, 2006.
- Gary C. McDonald. Ridge regression. Wiley Interdisciplinary Reviews: Computational Statistics, 1(1):93–100, 2009.
- Carl D. Meyer Jr. Generalized inversion of modified matrices. SIAM Journal on Applied Mathematics, 24(3):315–323, 1973.
- On model selection consistency of lasso for high-dimensional Ising models. In International Conference on Artificial Intelligence and Statistics, pages 6783–6805. PMLR, 2023.
- Ridge regression in practice. The American Statistician, 29(1):3–20, 1975.
- Definitions, methods, and applications in interpretable machine learning. Proceedings of the National Academy of Sciences, 116(44):22071–22080, 2019.
- Complexity analysis of the lasso regularization path. In Proceedings of the 29th International Coference on International Conference on Machine Learning, pages 1835–1842, 2012.
- A new approach to variable selection in least squares problems. IMA journal of numerical analysis, 20(3):389–403, 2000.
- On the LASSO and its dual. Journal of Computational and Graphical statistics, 9(2):319–337, 2000.
- Portfolio selection: A statistical learning approach. In Proceedings of the Third ACM International Conference on AI in Finance, pages 257–263, 2022.
- Anupam Prakash. Quantum algorithms for linear algebra and machine learning. University of California, Berkeley, 2014.
- John Preskill. Quantum computing in the NISQ era and beyond. Quantum, 2:79, 2018.
- Robust quantum minimum finding with an application to hypothesis selection. arXiv preprint arXiv:2003.11777, 2020.
- Interpretable machine learning: Fundamental principles and 10 grand challenges. Statistic Surveys, 16:1–85, 2022.
- Ridge Regression. Biased estimation for nonorthogonal problems. Technometrics, 12(1):55–67, 1970.
- Quantum support vector machine for big data classification. Physical review letters, 113(13):130503, 2014.
- Saharon Rosset. Following curved regularized optimization solution paths. Advances in Neural Information Processing Systems, 17, 2004.
- Volker Roth. The generalized lasso. IEEE transactions on neural networks, 15(1):16–28, 2004.
- Efficient feature size reduction via predictive forward selection. Pattern Recognition, 47(4):1664–1673, 2014.
- Saharon Rosset and Ji Zhu. Piecewise linear regularized solution paths. The Annals of Statistics, pages 1012–1030, 2007.
- SpaSM: A MATLAB toolbox for sparse statistical modeling. Journal of Statistical Software, 84:1–37, 2018.
- A sparse-group lasso. Journal of computational and graphical statistics, 22(2):231–245, 2013.
- Ridge regression learning algorithm in dual variables. Proceedings of the Fifteenth International Conference on Machine Learning, 1998.
- Prediction by linear regression on a quantum computer. Physical Review A, 94(2):022342, 2016.
- Mihailo Stojnic. A framework to characterize performance of LASSO algorithms. arXiv preprint arXiv:1303.7291, 2013.
- J.A.K. Suykens and J. Vandewalle. Least squares support vector machine classifiers. Neural processing letters, 9:293–300, 1999.
- Quantum regularized least squares solver with parameter estimate. Quantum Information Processing, 19:1–20, 2020.
- The contextual lasso: Sparse linear models via deep neural networks. arXiv preprint arXiv:2302.00878, 2023.
- A genetic algorithm-based method for feature subset selection. Soft Computing, 12:111–120, 2008.
- Robert Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society Series B: Statistical Methodology, 58(1):267–288, 1996.
- Ryan J. Tibshirani. The lasso problem and uniqueness. Electronic Journal of Statistics, 7:1456–1490, 2013.
- Joel A. Tropp. Just relax: Convex programming methods for identifying sparse signals in noise. IEEE transactions on information theory, 52(3):1030–1051, 2006.
- Sparsity and smoothness via the fused lasso. Journal of the Royal Statistical Society Series B: Statistical Methodology, 67(1):91–108, 2005.
- LASSO with cross-validation for genomic selection. Genetics research, 91(6):427–436, 2009.
- Hrishikesh D. Vinod. A survey of ridge regression and related techniques for improvements over ordinary least squares. The Review of Economics and Statistics, pages 121–131, 1978.
- Sequential forward feature selection with low computational cost. In 2005 13th European Signal Processing Conference, pages 1–4. IEEE, 2005.
- Wessel N. van Wieringen. Lecture notes on ridge regression. arXiv preprint arXiv:1509.09169, 2015.
- Martin J. Wainwright. Sharp thresholds for high-dimensional and noisy sparsity recovery using ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT-constrained quadratic programming (Lasso). IEEE transactions on information theory, 55(5):2183–2202, 2009.
- Martin J. Wainwright. High-dimensional statistics: A non-asymptotic viewpoint, volume 48. Cambridge university press, 2019.
- Guoming Wang. Quantum algorithm for linear regression. Physical review A, 96(1):012335, 2017.
- Feature subset selection and ranking for data dimensionality reduction. IEEE transactions on pattern analysis and machine intelligence, 29(1):162–166, 2006.
- Quantum algorithm for data fitting. Physical review letters, 109(5):050505, 2012.
- Genome-wide association analysis by lasso penalized logistic regression. Bioinformatics, 25(6):714–721, 2009.
- Unsupervised forward selection: a method for eliminating redundant variables. Journal of chemical information and computer sciences, 40(5):1160–1168, 2000.
- John Wright and Yi Ma. High-dimensional data analysis with low-dimensional models: Principles, computation, and applications. Cambridge University Press, 2022.
- Efficient generalized fused lasso and its applications. ACM Transactions on Intelligent Systems and Technology (TIST), 7(4):1–22, 2016.
- An improved quantum algorithm for ridge regression. IEEE Transactions on Knowledge and Data Engineering, 33(3):858–866, 2019.
- Algorithms for feature selection: An evaluation. In Proceedings of 13th international conference on pattern recognition, volume 2, pages 18–22. IEEE, 1996.
- Pathwise coordinate optimization for sparse learning: Algorithm and theory. The Annals of Statistics, 46(1):180 – 218, 2018.
- On model selection consistency of Lasso. The Journal of Machine Learning Research, 7:2541–2563, 2006.