Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Low-degree learning and the metric entropy of polynomials (2203.09659v4)

Published 17 Mar 2022 in cs.LG, cs.CC, math.CO, and stat.ML

Abstract: Let $\mathscr{F}{n,d}$ be the class of all functions $f:{-1,1}n\to[-1,1]$ on the $n$-dimensional discrete hypercube of degree at most $d$. In the first part of this paper, we prove that any (deterministic or randomized) algorithm which learns $\mathscr{F}{n,d}$ with $L_2$-accuracy $\varepsilon$ requires at least $\Omega((1-\sqrt{\varepsilon})2d\log n)$ queries for large enough $n$, thus establishing the sharpness as $n\to\infty$ of a recent upper bound of Eskenazis and Ivanisvili (2021). To do this, we show that the $L_2$-packing numbers $\mathsf{M}(\mathscr{F}{n,d},|\cdot|{L_2},\varepsilon)$ of the concept class $\mathscr{F}{n,d}$ satisfy the two-sided estimate $$c(1-\varepsilon)2d\log n \leq \log \mathsf{M}(\mathscr{F}{n,d},|\cdot|{L_2},\varepsilon) \leq \frac{2{Cd}\log n}{\varepsilon4}$$ for large enough $n$, where $c, C>0$ are universal constants. In the second part of the paper, we present a logarithmic upper bound for the randomized query complexity of classes of bounded approximate polynomials whose Fourier spectra are concentrated on few subsets. As an application, we prove new estimates for the number of random queries required to learn approximate juntas of a given degree, functions with rapidly decaying Fourier tails and constant depth circuits of given size. Finally, we obtain bounds for the number of queries required to learn the polynomial class $\mathscr{F}{n,d}$ without error in the query and random example models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Citations (9)

Summary

We haven't generated a summary for this paper yet.