Non-asymptotic state-evolution conjecture for quadratic networks (matrix compressed sensing)
Establish that, for over-parameterized two-layer quadratic neural networks trained with weight decay—equivalently, nuclear-norm regularized matrix compressed sensing—with any fixed λ>0 and Δ≥0, and for sufficiently large sample size n and dimension d, both the empirical risk minimizer’s excess risk and the Bayes-optimal risk concentrate within multiplicative o(1) around the deterministic quantities predicted by AMP state evolution fixed-point equations for these models.
Sponsor
References
Conjecture. Let \lambda>0, \Delta\geq0 and consider n,d\gg 1 sufficiently large. Then with a probability at least 1-o_n(1)-o_d(1), both the excess risk associated to the empirical risk minimizer (\ref{eq:def:quadratic_network}) and the Bayes-optimal risk satisfy |R(\hat{S}) - \mathsf R_{n,d}| = \mathsf R_{n,d}\cdot o_{n,d}(1).