Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
GPT-5.1
GPT-5.1 133 tok/s
Gemini 3.0 Pro 55 tok/s Pro
Gemini 2.5 Flash 164 tok/s Pro
Kimi K2 202 tok/s Pro
Claude Sonnet 4.5 39 tok/s Pro
2000 character limit reached

Sample Complexity Bounds

Updated 18 November 2025
  • Sample complexity bounds are quantitative measures defining the minimum number of data samples required to reach a target accuracy and confidence in various estimation and decision-making tasks.
  • They are applied across diverse domains such as PAC learning, reinforcement learning, convex optimization, and quantum measurement to provide finite-sample guarantees and reveal performance trade-offs.
  • By leveraging minimax and instance-dependent analyses along with structure-adapted techniques, these bounds guide the design of statistically efficient algorithms under differing conditions.

Sample complexity bounds quantify the minimal number of data samples required by an algorithm or estimator to achieve a specified level of accuracy and confidence in learning, system identification, estimation, or decision-making under uncertainty. These bounds are foundational in learning theory, statistics, signal processing, control, statistics, and quantum information, providing finite-sample guarantees and characterizing achievable performance. Cutting across classical PAC learning, margin-based classification, convex optimization, dynamical systems, reinforcement learning, quantum measurement learning, and modern high-dimensional models, sample complexity analysis reveals both minimax and instance-dependent rates, clarifies the trade-offs induced by structure (such as invariances, privacy, or model families), and leads to algorithmic designs optimized for statistical efficiency.

1. Foundational Definitions and Frameworks

Sample complexity bounds have diverse meanings across technical domains but are always anchored to three ingredients: a class of target objects (concepts, distributions, policies, measurements), an allowable error or loss metric (such as misclassification, total variation, prediction error, or regret), and a probability or confidence parameter.

  • PAC Learning (Realizable Case): Given concept class CC of VC-dimension dd, permissible error ϵ\epsilon, and confidence 1δ1-\delta, the sample complexity m(ϵ,δ)m(\epsilon,\delta) is the smallest mm such that a learner AA produces hh with Pr[err(h)ϵ]1δ\Pr[\mathrm{err}(h)\leq\epsilon] \geq 1-\delta for any cCc \in C. The optimal rate is m(ϵ,δ)=Θ(d+log(1/δ)ϵ)m(\epsilon,\delta) = \Theta\left( \frac{d+\log(1/\delta)}{\epsilon} \right) (Hanneke, 2015).
  • Margin-based Classification: For large-margin (e.g., SVM-type) classification with margin parameter γ\gamma and distribution PP, the sample complexity is governed by the margin-adapted dimension dm(γ,P)d_m(\gamma,P), with m=O~((dm(γ,P)+log(1/δ))/ϵ2)m = \widetilde{O}\left((d_m(\gamma,P) + \log(1/\delta)) / \epsilon^2\right) (Sabato et al., 2012).
  • Differential Privacy: In pure differentially private PAC learning, sample complexity SCDP(C)SCDP(C) for class CC is tightly characterized by randomized one-way public-coin communication complexity, and lower bounded by the Littlestone dimension: SCDP(C)=Ω(LDim(C))SCDP(C) = \Omega(LDim(C)), with possible gaps SCDP(C)VC(C)SCDP(C) \gg VC(C) (Feldman et al., 2014).
  • Convex Stochastic Programs: For SAA in convex stochastic programming, new bounds without metric entropy terms reach N=O(σ2/ϵ2)N = O(\sigma^2/\epsilon^2) under mild regularity and bypass linear dependence on dimension dd (Liu et al., 1 Jan 2024).
  • Quantum Learning: Quantum PAC sample complexity for measurement classes depends on the shadow-norm parameter VCV_{\mathcal{C}^*} and the size of extreme points; n=O(VClogC/ϵ2)n = O(V_{\mathcal{C}^*} \log |\mathcal{C}^*| / \epsilon^2), realizing the first non-linear scaling in the quantum setting (Heidari et al., 22 Aug 2024).

2. Classical Minimax and Instance-dependent Rates

Across learning, optimization, and decision-making, both worst-case (minimax) and instance-dependent rates arise.

  • Best-kk-Arm Identification: In multi-armed bandit selection, the minimal expected samples required to identify the top-kk arms is governed by instance-wise gap-dependent terms (H+H^+ and HH^-), which can be strictly smaller than the classical sum-of-inverse-gap-squared measure. Nearly instance-optimal algorithms achieve the lower bounds up to doubly-logarithmic factors (Chen et al., 2017).
  • Zero-Sum Matrix Games: Approximate Nash equilibrium identification in n×2n\times2 matrix games admits sample complexity lower bounds scaling as O(max{1/Δmin2,1/(ϵD)}log(1/δ))O(\max\{1/\Delta_{\min}^2,\,1/(\epsilon|D|)\}\log(1/\delta)), where Δmin\Delta_{\min} and DD are gap parameters of the payoff matrix. For favorable instances, exponentially fewer samples are needed (Maiti et al., 2023).
  • Reinforcement Learning—Actor-Critic: In infinite-horizon discounted MDPs, actor-critic methods for ϵ\epsilon-stationarity achieve sample complexity O((1γ)2ϵ2log(1/ϵ))O((1-\gamma)^{-2}\epsilon^{-2}\log(1/\epsilon)) (AC) and O((1γ)4ϵ3log(1/ϵ))O((1-\gamma)^{-4}\epsilon^{-3}\log(1/\epsilon)) (NAC), strictly better than policy gradient (PG) and natural policy gradient (NPG), where critic-based variance reduction is essential (Xu et al., 2020).
  • Diffusion Model Training: For high-dimensional deep generative models (e.g., diffusion models), end-to-end sample complexity scales only polylogarithmically in the inverse desired Wasserstein error, achieving m=poly(d)polylog(1/γ)m = {\rm poly}(d) \cdot \operatorname{polylog}(1/\gamma), closing a gap versus prior polynomial-in-1/γ1/\gamma rates (Gupta et al., 2023).
  • Constrained MDPs: For constrained average-reward MDPs, the minimax-optimal sample complexity is O~(SA(B+H)/(ϵ2ζ2))\widetilde{O}(SA(B+H)/(\epsilon^2 \zeta^2)) under strict feasibility; here S,AS,A are sizes, BB bounds transients, HH is the span, and ζ\zeta the Slater constant. This reveals a precise statistical penalty for feasibility constraints (Wei et al., 20 Sep 2025).

3. Structure-adapted and Distribution-dependent Bounds

Many sample complexity bounds leverage distributional or algebraic structure to obtain sharper results.

  • Distributional Dependence in Margin-based Learning: The margin-adapted dimension dm(γ,P)d_m(\gamma,P) exactly interpolates between worst-case (VC-dimension, Tr(Σ)/γ2\operatorname{Tr}(\Sigma)/\gamma^2) and favorable anisotropic regimes (effective dimension d\ll d), yielding tight sample complexity O(dm(γ,P)/ϵ2)O(d_m(\gamma,P)/\epsilon^2) for large-margin classification under sub-Gaussian product laws (Sabato et al., 2012).
  • Group Invariances: For distributions invariant under a group GG acting on a manifold MM, sample complexity for estimating Wasserstein, Sobolev, MMD, and L2L^2/ LL^\infty density divergences is reduced by a factor of G|G| (for finite GG) or by the normalized measure of the quotient M/GM/G (for continuous GG), and convergence rates in nn improve to match the "effective dimension" d=dim(M/G)d = \dim(M/G) (Tahmasebi et al., 2023).
  • Denoising and High-dimensional Estimation: In learning a simplex from noisy (additive Gaussian) samples in RK\mathbb{R}^K, the sample complexity is n=O~(K2/ϵ2)exp(Ω(K/SNR2))n = \widetilde{O}(K^2/\epsilon^2) \exp(\Omega(K/{\rm SNR}^2)), but with SNR K\gtrsim \sqrt{K}, the exponential phase vanishes, matching the noiseless case (Saberi et al., 2022).

4. Algorithmic and Information-theoretic Techniques

Sample complexity bounds are derived using a range of analytical and constructive tools, often leveraging information theory, combinatorial constructions, and algorithmic design.

  • Compression Schemes: For density estimation, the existence of data-dependent robust compression schemes directly yields tight upper bounds for mixtures and products (e.g., learning kk-mixtures of dd-dimensional Gaussians in TV requires Θ~(kd2/ϵ2)\widetilde{\Theta}(k d^2 / \epsilon^2) samples in the agnostic setting) (Ashtiani et al., 2017).
  • Information-theoretic Lower Bounds: Change-of-measure (likelihood ratio) arguments, Fano-type constructions, and KL-divergence sensitivity are used to prove lower bounds that often match upper bounds up to constants or logarithms (as in system identification (Jedra et al., 2019, Chatzikiriakos et al., 17 Sep 2024), margin learning (Sabato et al., 2012), bandit feedback (Chen et al., 2017)).
  • Finite-sample Monte Carlo: For Sequential Monte Carlo (SMC) schemes, the finite-sample complexity is determined by mixing properties, importance weight regularity, and the number of tempering stages, producing bounds within logarithmic factors of Markov chain Monte Carlo for log-concave and high-dimensional targets (Marion et al., 2018).
  • First-order and Gradient-based Methods in Bilevel and RL: Recent advances in bilevel RL circumvent Hessian dependence using the Polyak-Łojasiewicz property and penalty surrogates, attaining O(ϵ4)O(\epsilon^{-4}) rates in continuous state-action domains, improving dramatically over past O(ϵ6)O(\epsilon^{-6}) and higher orders (Gaur et al., 22 Mar 2025).

5. Separation Results and Limits

Situations where privacy, quantum structure, or other constraints fundamentally increase the sample complexity are precisely characterized.

  • Private vs. Non-private Learning: In PAC learning, pure (β=0\beta = 0) differential privacy can require exponentially more samples than both non-private or approximate (β>0\beta>0) DP; exact separations are constructed using representation dimension and communication complexity, with explicit classes achieving arbitrarily large SCDP/VC gaps (Feldman et al., 2014).
  • Quantum Learning vs. Classical: Quantum PAC sample complexity for measurement classes can scale linearly in C|\mathcal{C}| in the worst case, versus log-linear in the shadow-norm-exposed extreme points: n=O(VClogC)n = O(V_{\mathcal{C}^*} \log |\mathcal{C}^*|), restoring a near-classical logarithmic dependence under bounded-norm circumstances (Heidari et al., 22 Aug 2024).
  • Strict-constrained MDPs: Enforcing zero constraint violation in average-reward CMDPs requires 1/ζ21/\zeta^2 more samples compared to allowing ϵ\epsilon-approximate feasibility, reflecting a statistical hardness not present in unconstrained or discounted/finite-horizon analogues (Wei et al., 20 Sep 2025).

6. Parameter Dependence, Tightness, and Open Directions

Sample complexity bounds are sensitive to task-specific parameters, variance, model size, smoothness, problem structure, and required accuracy. In many cases, upper and lower bounds match up to logarithmic or constant factors; in others, subtle gaps remain.

Setting Core Sample Complexity Key Parameters Tightness/Separations
PAC-realizable (Hanneke, 2015) Θ((d+log(1/δ))/ϵ)\Theta((d+\log(1/\delta))/\epsilon) VC-dim dd, error/conf Tight, improved constant, no log(1/ϵ)\log(1/\epsilon)
Margin-based (Sabato et al., 2012) O~(dm(γ,P)/ϵ2)\widetilde{O}(d_m(\gamma,P)/\epsilon^2) Margin, covariance Tight both sides for sub-Gaussians
Private PAC (Feldman et al., 2014) Ω(LDim(C))\Omega(LDim(C)) (pure DP) LDim(C)LDim(C) vs VC(C)VC(C) Unbounded gap possible, separation pure/approximate DP
RNN regression (Akpinar et al., 2019) O~(a4b/ϵ2)\widetilde{O}(a^4 b/\epsilon^2) width, input length Deterministic upper bound
Influence maximization (Sadeh et al., 2019) O(sτϵ2log(n/δ))O(s \tau \epsilon^{-2} \log(n/\delta)) seed size, diffusion steps Removal of nn-factor, near-optimal
CMDP strict feasibility (Wei et al., 20 Sep 2025) O~(SA(B+H)/(ϵ2ζ2))\widetilde{O}(SA(B+H)/(\epsilon^2 \zeta^2)) problem size, slack Minimax tight for CAMDPs

Many open problems persist: refining constants in optimal PAC bounds (Hanneke, 2015), closing remaining logarithmic gaps in convex stochastic programming and SAA (Liu et al., 1 Jan 2024), extending metric-entropy-free rates to nonconvex or non-Lipschitz settings, generalization to more complex reinforcement learning and stochastic control models, and fully characterizing quantum-classical separations in high-dimensional limit regimes.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Sample Complexity Bounds.