Conjectured χ/ε sample complexity under the Factorization Hypothesis
Establish whether the sample complexity for learning the discrete conditional distribution p(y|x) with Kullback–Leibler (excess cross-entropy) loss at accuracy ε under the Factorization Hypothesis (FH: p(y|x) = ∏_{j=1}^{ℓ} p(y_j | pa_j) with input factorization X ≅ ∏_{i=1}^{k} X_i and output factorization Y ≅ ∏_{j=1}^{ℓ} Y_j, and parent sets I_j ⊆ [k]) is bounded by χ/ε, where χ = ∑_{j=1}^{ℓ} q_j × |pa_j|, q_j = |Y_j|, and |pa_j| = ∏_{i∈I_j} |X_i|.
Sponsor
References
Because the (hidden) factorization assumption transforms the learning task into \ell independent (but unknown) learning tasks, we conjecture the optimistic bound \chi / \ve on the sample complexity, where
\begin{equation}
\label{eq:SC}
\tag{SC}
\chi = \sum_{j=1}\ell q_j \times |\pa_j|, \end{equation}
which is always smaller than~eq:sc (and often {\em exponentially smaller}). The bound is optimistic because it is computed as if the factorization was known in advance; this can be heuristically justified by the fact that for most classical tasks in nonparametric statistics, the rate of estimation depends directly on structural assumptions (i.e. multi-index models, manifold hypothesis, etc) even when one does not know the precise instance of these assumptions (i.e. indices in the multi-index models, location of the manifold, etc).