Papers
Topics
Authors
Recent
2000 character limit reached

Equivalence of Approximate Message Passing and Low-Degree Polynomials in Rank-One Matrix Estimation (2212.06996v2)

Published 14 Dec 2022 in math.ST, math.PR, and stat.TH

Abstract: We consider the problem of estimating an unknown parameter vector ${\boldsymbol \theta}\in{\mathbb R}n$, given noisy observations ${\boldsymbol Y} = {\boldsymbol \theta}{\boldsymbol \theta}{\top}/\sqrt{n}+{\boldsymbol Z}$ of the rank-one matrix ${\boldsymbol \theta}{\boldsymbol \theta}{\top}$, where ${\boldsymbol Z}$ has independent Gaussian entries. When information is available about the distribution of the entries of ${\boldsymbol \theta}$, spectral methods are known to be strictly sub-optimal. Past work characterized the asymptotics of the accuracy achieved by the optimal estimator. However, no polynomial-time estimator is known that achieves this accuracy. It has been conjectured that this statistical-computation gap is fundamental, and moreover that the optimal accuracy achievable by polynomial-time estimators coincides with the accuracy achieved by certain approximate message passing (AMP) algorithms. We provide evidence towards this conjecture by proving that no estimator in the (broader) class of constant-degree polynomials can surpass AMP.

Citations (19)

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Paper to Video (Beta)

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube