Quantify the computational advantage of depth under gradient-based training
Establish a rigorous quantitative characterization, in an analytically tractable setting, of the computational advantage of deep neural networks trained with gradient-based methods relative to shallow models, specifying the criteria (e.g., sample complexity or generalization performance) by which the advantage is measured and the assumptions under which it holds.
References
A fundamental open problem is thus: Can one quantify the computational advantage of deep models trained with gradient-based methods with respect to shallow models in some analyzable setting?
— The Computational Advantage of Depth: Learning High-Dimensional Hierarchical Functions with Gradient Descent
(2502.13961 - Dandi et al., 19 Feb 2025) in Introduction (opening section)