Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 189 tok/s
Gemini 2.5 Pro 53 tok/s Pro
GPT-5 Medium 36 tok/s Pro
GPT-5 High 36 tok/s Pro
GPT-4o 75 tok/s Pro
Kimi K2 160 tok/s Pro
GPT OSS 120B 443 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

B-Calibration in Fixed-b Resampling

Updated 11 November 2025
  • B-Calibration is a method that calibrates resampling-based p-values under fixed-b asymptotics, addressing the impact of non-negligible block sizes.
  • It uses the limiting distribution of p-values—obtained via analytic or Monte Carlo methods—to correct nominal test levels and ensure valid coverage.
  • The approach extends to multivariate and functional inference, robustly calibrating confidence regions for scalar, vector, and infinite-dimensional parameters.

B-calibration refers to a set of statistical procedures and methodologies that address the calibration of resampling-based inference when the bandwidth parameter (block size, subsample length, or more generally, a smoothing or dependence parameter) is not asymptotically negligible compared to the sample size—so-called "fixed-b" asymptotics. The term particularly encapsulates the process of calibrating inference via the distribution of p-values under these fixed-b limits, thereby ensuring that statistical tests and confidence sets achieve accurate finite-sample coverage in dependent data settings such as time series. In the broader literature, "B-calibration" may also appear in the context of tuning parameters or performance calibration in high energy physics; however, the canonical usage for statistical inference is as rigorous methodology for p-value calibration under fixed-b block resampling (Shao et al., 2012).

1. Foundations: Fixed-b Asymptotics in Resampling

The classical resampling framework for dependent data (e.g., time series) involves creating either subsamples or blocks of length lnl_n, with the relative size bn=ln/nb_n = l_n / n. Traditionally, asymptotics are developed under the regime bn0b_n \to 0 as nn \to \infty ("small-b"), ensuring the influence of the bandwidth parameter vanishes at first order. Standard plug-in confidence sets or tests then remain pivotal in their limiting properties, and the bandwidth selection primarily only affects higher-order error.

In contrast, fixed-b asymptotics assume bnb(0,1]b_n \to b \in (0,1] as nn\to\infty, so the block size remains a constant positive fraction of the sample size. This change in theoretical regime alters the limiting distribution of key statistics—most notably, the distribution of p-values under the null hypothesis—such that they are no longer uniform but instead explicit bb-dependent distributions. This shift necessitates calibration to recover accurate test levels and interval coverages (Shao et al., 2012).

2. Limiting Distribution of Resampling-Based p-values

For a statistic Tn(X1,,Xn)T_n(X_1, \dots, X_n) targeting a parameter θ\theta (with, e.g., n(Tnθ)\sqrt{n}(T_n - \theta) converging to a nondegenerate limit), the resampling-based (subsampling or block bootstrap) p-value, typically

pn(l)=1Nj=1N1{Tj,lTn},N=nl+1p_n(l) = \frac{1}{N} \sum_{j=1}^{N} \mathbf{1}\{T_{j,l} \geq T_n\}, \qquad N = n - l + 1

(where Tj,lT_{j,l} is TT computed on the jjth block/subsample of length ll), converges to a non-uniform law UbU_b under fixed-b asymptotics. That is, pn(l)Ubp_n(l) \Rightarrow U_b with cumulative distribution function Fb(u)=Pr{Ubu}F_b(u) = \Pr\{U_b \leq u\}.

For moving block bootstrap, the analogous p-value pn(l)p_n^*(l) also converges to a bb-dependent law UbU_b^*. The explicit forms of these laws are typically functionals of underlying Gaussian or Brownian motion processes and can be determined by simulation or analytic computation for given bb (Shao et al., 2012).

3. P-Value Calibration Procedure

Because the limiting p-value is non-uniform (and depends on the chosen bb), classical critical values (e.g., α\alpha for a level-α\alpha test) cannot guarantee actual rejection probability α\alpha. The calibration procedure replaces the nominal cutoff with the bb-quantile of the null distribution: cb(α)=inf{u[0,1]:Fb(u)α}c_b(\alpha) = \inf\{u \in [0,1]: F_b(u) \geq \alpha\} so that under the null,

Pr{pn(l)cb(α)}=α\Pr\left\{p_n(l) \leq c_b(\alpha)\right\} = \alpha

which restores correct test size. For two-sided confidence intervals or regions, this is equivalent to "inverting" the calibrated test: Cn(1α)={θ:pn(l;θ)cb(α)}C_n(1-\alpha) = \{\theta: p_n(l; \theta) \geq c_b(\alpha)\} This B-calibration paradigm converts the non-pivotal, resampling-based p-value into an exact pivotal test in the fixed-b regime (Shao et al., 2012).

Type Limiting p-value law Calibration Quantile
Subsample UbU_b cb(α)=Fb1(α)c_b(\alpha) = F_b^{-1}(\alpha)
Bootstrap UbU_b^* cb(α)=Fb1(α)c_b^*(\alpha) = F_b^{*-1}(\alpha)

4. Generalization to Multivariate and Functional Inference

The B-calibration methodology is not limited to scalar parameters. For vector-valued settings, with estimators θ^nRk\hat{\theta}_n \in \mathbb{R}^k and statistics such as n(θ^nθ)\|\sqrt{n}(\hat{\theta}_n - \theta)\|, the limiting p-value law Ub(k)U_b(k) is again non-uniform and depends on (b,k)(b, k). For infinite-dimensional functionals—e.g., empirical distribution functions or spectral distribution functions—the sup-norm statistic over a class T\mathcal{T} follows an analogous limiting distribution UbU_b^{\infty} under fixed-b, with corresponding quantile cb(α)c_b^\infty(\alpha) for confidence bands.

The calibration is then: Cn(1α)={θRk:pn(l;θ)cb,k(α)}C_n(1-\alpha) = \{\theta \in \mathbb{R}^k: p_n(l;\theta) \geq c_{b,k}(\alpha)\} for finite-dimensional θ\theta, and

Bn(1α)={Φ:pn(l;Φ)cb(α)}\mathcal{B}_n(1-\alpha) = \{\Phi: p_n(l; \Phi) \geq c_b^\infty(\alpha)\}

for functional inference (Shao et al., 2012).

5. Implementation and Practical Guidance

Implementation of B-calibration involves tabulating the relevant critical values cb(α)c_b(\alpha) for a set of bb values, either via analytical simulation of the limiting process or by large-scale Monte Carlo of block-resampled statistics. For applied work, the recommended approach is:

  • Simulate 5000+ paths of the relevant Gaussian process, compute the empirical bb-distribution of the resampling p-value, and record quantiles cb(α)c_b(\alpha).
  • Alternatively, use a secondary Monte Carlo based on the block bootstrap under the null to approximate the limit law.
  • Prepare a lookup table for cb(α)c_b(\alpha) across a range of bb values (e.g., b{0.02,0.05,0.1,0.15}b\in\{0.02, 0.05, 0.1, 0.15\}), which may be accurately modeled by low-order polynomials in bb.

Block length selection remains a balance between bias (short blocks) and variance (long blocks), but the fixed-b theory provides explicit quantification of bb-dependence, reducing sensitivity to block-length choice when calibrated (Shao et al., 2012).

6. Empirical Performance and Coverage Properties

Extensive simulation in (Shao et al., 2012) demonstrates that traditional small-b methods systematically undercover, especially with strong dependence. Fixed-b B-calibrated intervals exhibit nominal coverage closer to the target 1α1-\alpha across a range of settings. The increased interval or band width is typically modest (10–20%), representing a tradeoff for improved statistical accuracy. Performance gains extend to confidence regions and bands for vector and infinite-dimensional targets, provided the distribution of the resampling p-value is properly approximated.

7. Applicability and Impact in Statistical Inference

B-calibration provides a unified and formally justified approach for accurate inference in time series and other dependent data contexts, incorporating the real-world influence of the bandwidth parameter at leading order. This directly improves confidence set coverage, test validity, and the reliability of bootstrap/subsampling-based procedures. Its generality covers a wide spectrum of practical estimation problems, including but not limited to means, distribution functions, spectral densities, and beyond, ensuring robust results without reliance on asymptotically negligible block sizes.

The methodology has been adopted across advanced empirical studies where resampling is essential and the classical small-b assumption is either theoretically invalid or practically difficult to maintain. It offers a lucid quantification and correction mechanism for the resampling bias introduced by non-negligible bandwidth choices, thus refining the inferential power of modern statistics (Shao et al., 2012).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to B-Calibration.