Papers
Topics
Authors
Recent
Search
2000 character limit reached

Analysis of the rate of convergence of an over-parametrized deep neural network estimate learned by gradient descent

Published 4 Oct 2022 in math.ST and stat.TH | (2210.01443v1)

Abstract: Estimation of a regression function from independent and identically distributed random variables is considered. The $L_2$ error with integration with respect to the design measure is used as an error criterion. Over-parametrized deep neural network estimates are defined where all the weights are learned by the gradient descent. It is shown that the expected $L_2$ error of these estimates converges to zero with the rate close to $n{-1/(1+d)}$ in case that the regression function is H\"older smooth with H\"older exponent $p \in [1/2,1]$. In case of an interaction model where the regression function is assumed to be a sum of H\"older smooth functions where each of the functions depends only on $d*$ many of $d$ components of the design variable, it is shown that these estimates achieve the corresponding $d*$-dimensional rate of convergence.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.