Optimal High-Probability PAC Bound for Majority-of-Three
Establish that the Majority-of-Three algorithm—defined as partitioning the training sample S into three equal-sized disjoint subsets S1, S2, S3, running the same empirical risk minimization (ERM) algorithm on each labeled subset, and returning the majority vote Maj(f̂_{S1}, f̂_{S2}, f̂_{S3})—achieves the optimal high-probability error bound O(d/n + (1/n) log(1/δ)) for every binary concept class F ⊆ {0,1}^X of Vapnik–Chervonenkis (VC) dimension d, every distribution P over X, every target function f* ∈ F, any ERM implementation, any sample size n, and all confidence parameters δ ∈ (0,1).
Sponsor
References
Because of this, we conjecture that Majority-of-Three is in fact optimal for all δ and leave this as an open question for future research.
— Majority-of-Three: The Simplest Optimal Learner?
(2403.08831 - Aden-Ali et al., 12 Mar 2024) in Introduction, Subsection “The simplest possible optimal algorithm?”, immediately following Theorem \ref{highprobboundsection:theorem}