Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
129 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Risk Bounds for High-dimensional Ridge Function Combinations Including Neural Networks (1607.01434v4)

Published 5 Jul 2016 in math.ST, stat.ML, and stat.TH

Abstract: Let $ f{\star} $ be a function on $ \mathbb{R}d $ with an assumption of a spectral norm $ v_{f{\star}} $. For various noise settings, we show that $ \mathbb{E}|\hat{f} - f{\star} |2 \leq \left(v4_{f{\star}}\frac{\log d}{n}\right){1/3} $, where $ n $ is the sample size and $ \hat{f} $ is either a penalized least squares estimator or a greedily obtained version of such using linear combinations of sinusoidal, sigmoidal, ramp, ramp-squared or other smooth ridge functions. The candidate fits may be chosen from a continuum of functions, thus avoiding the rigidity of discretizations of the parameter space. On the other hand, if the candidate fits are chosen from a discretization, we show that $ \mathbb{E}|\hat{f} - f{\star} |2 \leq \left(v3_{f{\star}}\frac{\log d}{n}\right){2/5} $. This work bridges non-linear and non-parametric function estimation and includes single-hidden layer nets. Unlike past theory for such settings, our bound shows that the risk is small even when the input dimension $ d $ of an infinite-dimensional parameterized dictionary is much larger than the available sample size. When the dimension is larger than the cube root of the sample size, this quantity is seen to improve the more familiar risk bound of $ v_{f{\star}}\left(\frac{d\log (n/d)}{n}\right){1/2} $, also investigated here.

Citations (68)

Summary

We haven't generated a summary for this paper yet.