Non-asymptotic state-evolution conjecture for diagonal networks (LASSO)
Establish that, for diagonal two-layer linear networks with ℓ2 weight decay—equivalently, LASSO regression—with any fixed λ>0 and Δ≥0, and for sufficiently large sample size n and dimension d, both the empirical risk minimizer’s excess risk and the Bayes-optimal risk concentrate within multiplicative o(1) around the deterministic quantities predicted by AMP state evolution fixed-point equations for LASSO.
References
Conjecture. Let \lambda>0, \Delta\geq0 and consider n,d\gg 1 sufficiently large. Then with a probability at least 1-o_n(1)-o_d(1), both the excess risk associated to the empirical risk minimizer (\ref{eq:def:lasso}) and the Bayes-optimal risk satisfy |R(\hat{\theta}) - \mathsf R_{n,d}| = \mathsf R_{n,d}\cdot o_{n,d}(1).
— Scaling Laws and Spectra of Shallow Neural Networks in the Feature Learning Regime
(2509.24882 - Defilippis et al., 29 Sep 2025) in Section Non-asymptotic state evolution, Conjecture (following the quadratic-network conjecture)