2000 character limit reached
Spurious Local Minima of Deep ReLU Neural Networks in the Neural Tangent Kernel Regime (1806.04884v3)
Published 13 Jun 2018 in stat.ML and cs.LG
Abstract: In this paper, we theoretically prove that the deep ReLU neural networks do not lie in spurious local minima in the loss landscape under the Neural Tangent Kernel (NTK) regime, that is, in the gradient descent training dynamics of the deep ReLU neural networks whose parameters are initialized by a normal distribution in the limit as the widths of the hidden layers tend to infinity.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.