Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Estimation of a function of low local dimensionality by deep neural networks (1908.11140v3)

Published 29 Aug 2019 in stat.ML, cs.LG, math.ST, and stat.TH

Abstract: Deep neural networks (DNNs) achieve impressive results for complicated tasks like object detection on images and speech recognition. Motivated by this practical success, there is now a strong interest in showing good theoretical properties of DNNs. To describe for which tasks DNNs perform well and when they fail, it is a key challenge to understand their performance. The aim of this paper is to contribute to the current statistical theory of DNNs. We apply DNNs on high dimensional data and we show that the least squares regression estimates using DNNs are able to achieve dimensionality reduction in case that the regression function has locally low dimensionality. Consequently, the rate of convergence of the estimate does not depend on its input dimension $d$, but on its local dimension $d*$ and the DNNs are able to circumvent the curse of dimensionality in case that $d*$ is much smaller than $d$. In our simulation study we provide numerical experiments to support our theoretical result and we compare our estimate with other conventional nonparametric regression estimates. The performance of our estimates is also validated in experiments with real data.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Michael Kohler (23 papers)
  2. Adam Krzyzak (9 papers)
  3. Sophie Langer (14 papers)
Citations (7)

Summary

We haven't generated a summary for this paper yet.