Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Low Rank Matrix Completion via Robust Alternating Minimization in Nearly Linear Time (2302.11068v3)

Published 21 Feb 2023 in cs.LG, cs.DS, math.OC, and stat.ML

Abstract: Given a matrix $M\in \mathbb{R}{m\times n}$, the low rank matrix completion problem asks us to find a rank-$k$ approximation of $M$ as $UV\top$ for $U\in \mathbb{R}{m\times k}$ and $V\in \mathbb{R}{n\times k}$ by only observing a few entries specified by a set of entries $\Omega\subseteq [m]\times [n]$. In particular, we examine an approach that is widely used in practice -- the alternating minimization framework. Jain, Netrapalli, and Sanghavi [JNS13] showed that if $M$ has incoherent rows and columns, then alternating minimization provably recovers the matrix $M$ by observing a nearly linear in $n$ number of entries. While the sample complexity has been subsequently improved [GLZ17], alternating minimization steps are required to be computed exactly. This hinders the development of more efficient algorithms and fails to depict the practical implementation of alternating minimization, where the updates are usually performed approximately in favor of efficiency. In this paper, we take a major step towards a more efficient and error-robust alternating minimization framework. To this end, we develop an analytical framework for alternating minimization that can tolerate a moderate amount of errors caused by approximate updates. Moreover, our algorithm runs in time $\widetilde O(|\Omega| k)$, which is nearly linear in the time to verify the solution while preserving the sample complexity. This improves upon all prior known alternating minimization approaches which require $\widetilde O(|\Omega| k2)$ time.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (64)
  1. Blendenpik: Supercharging lapack’s least-squares solver. SIAM J. Sci. Comput., 2010.
  2. Optimal cur matrix decompositions. In Proceedings of the forty-sixth annual ACM symposium on Theory of computing (STOC), pages 353–362, 2014.
  3. A near-optimal algorithm for approximating the john ellipsoid. In Conference on Learning Theory, pages 849–873. PMLR, 2019.
  4. Nearly optimal robust matrix completion. In Proceedings of the 34th International Conference on Machine Learning - Volume 70, ICML’17, page 797–805, 2017.
  5. Robust principal component analysis? J. ACM, 2011.
  6. Exact matrix completion via convex optimization. Commun. ACM, 2012.
  7. The power of convex relaxation: Near-optimal matrix completion. IEEE Trans. Inf. Theor., 2010.
  8. Low-rank approximation and regression in input sparsity time. In STOC, 2013.
  9. Optimal sketching for kronecker product regression and low rank approximation. Advances in neural information processing systems, 32, 2019.
  10. Sketching for kronecker product regression and p-splines. In International Conference on Artificial Intelligence and Statistics, pages 1299–1308. PMLR, 2018.
  11. Faster matrix multiplication via asymmetric hashing. In FOCS, 2023.
  12. Noisy matrix completion using alternating minimization. In Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2013, Prague, Czech Republic, September 23-27, 2013, Proceedings, Part II 13, pages 194–209. Springer, 2013.
  13. On traffic matrix completion in the internet. In Proceedings of the 2012 Internet Measurement Conference, IMC ’12, page 399–412, New York, NY, USA, 2012. Association for Computing Machinery.
  14. Matrix completion from O⁢(n)𝑂𝑛{O}(n)italic_O ( italic_n ) samples in linear time. In Satyen Kale and Ohad Shamir, editors, Proceedings of the 2017 Conference on Learning Theory, volume 65 of Proceedings of Machine Learning Research, pages 940–947. PMLR, 07–10 Jul 2017.
  15. Moritz Hardt. Understanding alternating minimization for matrix completion. In 2014 IEEE 55th Annual Symposium on Foundations of Computer Science, pages 651–660. IEEE, 2014.
  16. Solving sdp faster: A robust ipm framework and efficient implementation. In FOCS, 2022.
  17. Computational limits for matrix completion. In Conference on Learning Theory, pages 703–725. PMLR, 2014.
  18. Fast matrix completion without the condition number. In Conference on learning theory, pages 638–678. PMLR, 2014.
  19. A faster interior point method for semidefinite programming. In FOCS, 2020.
  20. Fast exact matrix completion with finite samples. In Peter Grünwald, Elad Hazan, and Satyen Kale, editors, Proceedings of The 28th Conference on Learning Theory, volume 40 of Proceedings of Machine Learning Research, pages 1007–1034, Paris, France, 03–06 Jul 2015. PMLR.
  21. Low-rank matrix completion using alternating minimization. In Proceedings of the forty-fifth annual ACM symposium on Theory of computing, pages 665–674, 2013.
  22. Charles R Johnson. Matrix completion problems: a survey. In Matrix theory and applications, volume 40, pages 171–198, 1990.
  23. Matrix factorization techniques for recommender systems. Computer, 42(8):30–37, 2009.
  24. Matrix completion in almost-verification time. In 2023 IEEE 64th Annual Symposium on Foundations of Computer Science, FOCS’23, 2023.
  25. Semi-random sparse recovery in nearly linear time. In Conference on Learning Theory, COLT’23, 2023.
  26. Sampling methods for the nystrom method. Journal of Machine Learning Research, 2012.
  27. Matrix completion from a few entries. In 2009 IEEE International Symposium on Information Theory, 2009.
  28. Yehuda Koren. The bellkor solution to the netflix grand prize. Netflix prize documentation, 81(2009):1–10, 2009.
  29. Faster ridge regression via the subsampled randomized hadamard transform. In Advances in neural information processing systems (NIPS), pages 369–377, 2013.
  30. The geometry of graphs and some of its algorithmic applications. In Proceedings 35th Annual Symposium on Foundations of Computer Science, 1994.
  31. Recovery guarantee of weighted low-rank approximation via alternating minimization. In International Conference on Machine Learning, pages 2358–2367. PMLR, 2016.
  32. A new theory for matrix completion. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017.
  33. Matrix completion using alternating minimization for distribution system state estimation. In 2020 IEEE International Conference on Communications, Control, and Computing Technologies for Smart Grids (SmartGridComm), 2020.
  34. Solving empirical risk minimization in the current matrix multiplication time. In Conference on Learning Theory, pages 2140–2157. PMLR, 2019.
  35. On convergence of the alternating projection method for matrix completion and sparse recovery problems, 2017.
  36. Randomized block krylov methods for stronger and faster approximate singular value decomposition. In Advances in Neural Information Processing Systems, volume 2015-January, pages 1396–1404, 2015.
  37. Lsrn: A parallel iterative solver for strongly over- or underdetermined systems. SIAM Journal on Scientific Computing, 36(2):C95–C118, 2014.
  38. Can matrix coherence be efficiently and accurately estimated? In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, Proceedings of Machine Learning Research, 2011.
  39. The optimal perturbation bounds of the moore–penrose inverse under the frobenius norm. Linear algebra and its applications, 432(4):956–963, 2010.
  40. Yu E Nesterov. A method for solving the convex programming problem with convergence rate O⁢(1k2)𝑂1superscript𝑘2{O}(\frac{1}{k^{2}})italic_O ( divide start_ARG 1 end_ARG start_ARG italic_k start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT end_ARG ). In Dokl. akad. nauk Sssr, volume 269, pages 543–547, 1983.
  41. Low-rank matrix completion: A contemporary survey. IEEE Access, 7:94215–94237, 2019.
  42. Osnap: Faster numerical linear algebra algorithms via sparser subspace embeddings. In 2013 ieee 54th annual symposium on foundations of computer science, pages 117–126. IEEE, 2013.
  43. Fast regression with an ℓ∞subscriptℓ\ell_{\infty}roman_ℓ start_POSTSUBSCRIPT ∞ end_POSTSUBSCRIPT guarantee. In ICALP, 2017.
  44. An online and unified algorithm for projection matrix vector multiplication with application to empirical risk minimization. In AISTATS, 2023.
  45. Benjamin Recht. A simpler approach to matrix completion. J. Mach. Learn. Res., 12:3413–3430, dec 2011.
  46. Jasson D. M. Rennie and Nathan Srebro. Fast maximum margin matrix factorization for collaborative prediction. In Proceedings of the 22nd International Conference on Machine Learning, ICML ’05, page 713–719, New York, NY, USA, 2005. Association for Computing Machinery.
  47. Weighted low rank approximations with provable guarantees. In Proceedings of the forty-eighth annual ACM symposium on Theory of Computing, pages 250–263, 2016.
  48. Dynamic tensor product regression. In NeurIPS, 2022.
  49. Guaranteed matrix completion via non-convex factorization. IEEE Transactions on Information Theory, 62(11):6535–6579, 2016.
  50. Graph sparsification by effective resistances. SIAM Journal on Computing, 40(6):1913–1926, 2011.
  51. Fast sketching of polynomial kernels of polynomial degree. In International Conference on Machine Learning, pages 9812–9823. PMLR, 2021.
  52. Low rank approximation with entrywise l1-norm error. In Proceedings of the 49th Annual ACM SIGACT Symposium on Theory of Computing, pages 688–701, 2017.
  53. Average case column subset selection for entrywise ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT-norm loss. Advances in Neural Information Processing Systems, 32, 2019.
  54. Towards a zero-one law for column subset selection. Advances in Neural Information Processing Systems, 32, 2019.
  55. Relative error tensor low rank approximation. In Proceedings of the Thirtieth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 2772–2789. SIAM, 2019.
  56. Anthony Man-Cho So and Yinyu Ye. Theory of semidefinite programming for sensor network localization. In Proceedings of the Sixteenth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA ’05, page 405–414, USA, 2005. Society for Industrial and Applied Mathematics.
  57. Oblivious sketching-based central path method for linear programming. In International Conference on Machine Learning, pages 9835–9847. PMLR, 2021.
  58. Joel A. Tropp. Improved analysis of the subsampled randomized hadamard transform. Adv. Data Sci. Adapt. Anal., 3, 2011.
  59. Per-Åke Wedin. Perturbation theory for pseudo-inverses. BIT Numerical Mathematics, 13(2):217–232, 1973.
  60. H. Weyl. Das asymptotische verteilungsgesetz der eigenwerte linearer partieller differentialgleichungen (mit einer anwendung auf die theorie der hohlraumstrahlung). Mathematische Annalen, 71:441–479, 1912.
  61. Breaking the linear iteration cost barrier for some well-known conditional gradient methods using maxip data-structures. Advances in Neural Information Processing Systems, 34:5576–5589, 2021.
  62. A tale of two efficient value iteration algorithms for solving linear mdps with large action space. In AISTATS, 2023.
  63. Angles between subspaces and their tangents. Journal of Numerical Mathematics, 21(4):325–340, 2013.
  64. A nonconvex optimization framework for low rank matrix estimation. In C. Cortes, N. Lawrence, D. Lee, M. Sugiyama, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 28. Curran Associates, Inc., 2015.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Yuzhou Gu (25 papers)
  2. Zhao Song (253 papers)
  3. Junze Yin (26 papers)
  4. Lichen Zhang (21 papers)
Citations (19)

Summary

We haven't generated a summary for this paper yet.