Papers
Topics
Authors
Recent
Search
2000 character limit reached

Equivalent Sample Size in Bayesian Models

Updated 26 January 2026
  • Equivalent sample size is a measure that translates the information in Bayesian priors or posteriors into the scale of independent sample counts.
  • It is used in Bayesian network scoring and experimental design to calibrate prior strength, balance model complexity, and inform sensitivity analyses.
  • Recent estimation techniques, such as nonparametric regression methods, provide rapid and precise ESS calculations in complex, nonconjugate scenarios.

Equivalent sample size (ESS), also known as effective sample size or prior effective sample size, is a quantitative measure expressing the information content of a probability distribution—usually a Bayesian prior or posterior—in terms of the hypothetical number of independent observations that would provide equivalent information under the corresponding likelihood model. ESS plays a central role in Bayesian modeling, model selection, and experimental design, enabling clear communication and calibration of prior assumptions, especially in scenarios where external or historical information is incorporated.

1. Foundations and Definitions

ESS formalizes the translation of prior or posterior information into the scale of sample size. For a scalar parameter θ\theta, the ESS of a prior distribution %%%%1%%%% is defined as the number n0n_0 such that the prior conveys as much Fisher-information as n0n_0 real data points under the model, i.e.,

ESS=n0whereEθ[IF(θ)]n0=information in p(θ)ESS = n_0 \quad \text{where} \quad \mathbb{E}_{\theta} \left[ \mathcal{I}_F(\theta) \right] n_0 = \text{information in } p(\theta)

Here, IF(θ)\mathcal{I}_F(\theta) is the expected Fisher-information of one new data point at θ\theta.

In Bayesian–Dirichlet equivalent uniform (BDeu) scoring for discrete Bayesian networks, the ESS parameter α\alpha directly controls the Dirichlet prior’s strength: for each node ViV_i and conditioning configuration, the Dirichlet’s “pseudo-counts” are set as α/(riqi)\alpha/(r_i q_i), where rir_i is the variable’s arity and qiq_i is the number of parent configurations (Silander et al., 2012).

ESS also emerges in study design and evaluation, quantifying how prior or external data alters precision and power calculations, e.g., by allowing formal sample size reduction in clinical or equivalence trials (Neuenschwander et al., 2019).

2. ESS in Bayesian Network Scoring

In Bayesian network structure learning, particularly under the BDeu marginal likelihood, ESS is parametrized as the scalar α\alpha. The BDeu score for network GG given data DD is

BDeu(GD,α)=i=1nj=1qiΓ(αij)Γ(Nij+αij)k=1riΓ(Nijk+αijk)Γ(αijk)\mathrm{BDeu}(G|D, \alpha) = \prod_{i=1}^n \prod_{j=1}^{q_i} \frac{\Gamma(\alpha_{ij})}{\Gamma(N_{ij} + \alpha_{ij})} \prod_{k=1}^{r_i} \frac{\Gamma(N_{ijk} + \alpha_{ijk})}{\Gamma(\alpha_{ijk})}

with

αijk=αriqi,αij=k=1riαijk=αqi\alpha_{ijk} = \frac{\alpha}{r_i q_i}, \qquad \alpha_{ij} = \sum_{k=1}^{r_i} \alpha_{ijk} = \frac{\alpha}{q_i}

where NijkN_{ijk} counts data configurations as described in (Silander et al., 2012).

The parameter α\alpha tunes the trade-off between data likelihood and prior-induced structure penalty. Larger α\alpha reduces the complexity penalty and pushes the maximum a posteriori (MAP) structure toward denser graphs, while smaller α\alpha heavily penalizes additional edges unless the data splits are highly informative.

Empirically, even small changes in α\alpha can lead to different MAP structures. Across benchmark datasets, the number of arcs in the optimal network—and thus the encoded independence structure—varies substantially as α\alpha is tuned, often with each plausible value within a broad range producing a distinct model (Silander et al., 2012).

3. ESS Calculation and Methodological Variants

Several approaches to quantifying ESS have been developed, especially for general Bayesian models where the prior is not conjugate:

  • Variance-ratio (VR) ESS: ESSVR=Eθ[IF(θ)1]/Var[θ]\mathrm{ESS}_{\mathrm{VR}} = \mathbb{E}_\theta[\mathcal{I}_F(\theta)^{-1}]/\operatorname{Var}[\theta]
  • Precision-ratio (PR) ESS: ESSPR=Var[θ]1/Eθ[IF(θ)]\mathrm{ESS}_{\mathrm{PR}} = \operatorname{Var}[\theta]^{-1}/\mathbb{E}_\theta[\mathcal{I}_F(\theta)]
  • Morita-Thall-Müller (MTM) ESS: Based on local information at the prior mean, [i(p)(θˉ)i(p0)(θˉ)]/EY1θˉ[IF(Y1;θˉ)][i(p)(\bar{\theta}) - i(p_0)(\bar{\theta})] / \mathbb{E}_{Y_1 | \bar{\theta}}[\mathcal{I}_F(Y_1; \bar{\theta})]
  • Expected local-information-ratio (ELIR) ESS: ESSELIR=Eθ[i(p)(θ)IF(θ)]\mathrm{ESS}_{\mathrm{ELIR}} = \mathbb{E}_\theta\left[\frac{i(p)(\theta)}{\mathcal{I}_F(\theta)}\right] where i(p)(θ)=d2dθ2logp(θ)i(p)(\theta) = -\frac{d^2}{d\theta^2} \log p(\theta)

In one-parameter exponential families, most definitions agree, but in nonconjugate models these methods can diverge substantially due to global/local information mismatch, parameter dependence in curvature, or differences between mean and mode anchoring (Neuenschwander et al., 2019).

The ELIR definition uniquely satisfies predictive consistency, i.e., after observing NN new data points, the expected posterior ESS increases by NN: EYN[ESSpost(YN)]=ESSprior+NE_{Y_N}[ESS_{\mathrm{post}}(Y_N)] = ESS_{\mathrm{prior}} + N (Neuenschwander et al., 2019). This property makes ELIR a rigorous metric for both prior and posterior ESS.

4. ESS Estimation Techniques

ESS can be analytically computed in simple, conjugate cases, but more generally requires estimation, especially for complex or nonstandard priors.

In Gaussian approximation methods for expected value of sample information (EVSI), the prior on parameter ϕ\phi is taken as approximately normal with variance σ2/n0\sigma^2/n_0. Then, n0=σ2/Var[ϕ]n_0 = \sigma^2/\operatorname{Var}[\phi]. More generally, variance-ratio and regression-based estimators are used:

n^0=n(Varϕ[ϕ]VarXn[E(ϕXn)]1)\hat n_0 = n \left( \frac{\operatorname{Var}_\phi[\phi]}{\operatorname{Var}_{X_n}[E(\phi|X_n)]} - 1 \right)

n^0=n(VarXn[Xˉn]Varϕ[ϕ]1)\hat n_0 = n \left( \frac{\operatorname{Var}_{X_n}[\bar X_n]}{\operatorname{Var}_\phi[\phi]} - 1 \right)

where nn is the simulated future sample size (Li et al., 2024).

A notable recent advance is nonparametric regression-based ESS estimation. Simulated draws (ϕm,Xnm)(\phi^m, X_n^m) are generated, a regression model ϕmg(T(Xnm))\phi^m \approx g(T(X_n^m)) is fit (with T()T(\cdot) a sufficient statistic or informative summary), and ESS is estimated by the variance ratio of ϕm\phi^m to fitted values g^(Tm)\hat g(T^m). This approach matches analytic results to within 1% in classic cases, and significantly reduces computation time compared to MCMC-based methods (Li et al., 2024).

Assumptions include well-approximated normality and informativeness of the summary T(Xn)T(X_n). The method is primarily restricted to univariate ϕ\phi; generalizations to multivariate settings and more flexible summaries are areas of ongoing research.

5. Sensitivity and Interpretation of ESS

The information encoded by ESS is highly sensitive to methodological details. In Bayesian network learning, small α\alpha variations can flip individual arcs or, more generally, produce markedly different independence structures. Empirical findings on dozens of datasets confirm that the number of distinct MAP structures realized across a moderate α\alpha interval (e.g., [1,100][1,100]) can be large, justifying the view that ESS is not only a quantifier of prior strength but also a determinant of model complexity (Silander et al., 2012).

Because ESS influences the trade-off between fit and complexity, arbitrary selection (e.g., always setting α=1\alpha=1) is discouraged. Principled strategies involve:

  • Integrating over α\alpha with a prior (e.g., uniform on [1,100][1,100]), maximizing the α\alpha-integrated posterior for the MAP model, or
  • Choosing (G,α)(G^*, \alpha^*) to maximize the marginal likelihood jointly (prequential approach)

In all methods evaluated, the optimal α\alpha typically lies near the average variable arity in the Bayesian network, and integrating out α\alpha or maximizing likelihood leads to consonant model choices (Silander et al., 2012).

6. Applications in Design, Evaluation, and Meta-Analysis

In trial design, prior ESS quantifies the “borrowed” information from historical or external controls. Knowledge of n0n_0 can be used to:

  • Reduce internal sample size without loss of information,
  • Calibrate type I/II error rates and power, and
  • Evaluate the contribution of hierarchical or mixture priors in predictive subgroup analyses (Neuenschwander et al., 2019).

In hierarchical analyses, posterior ESS can also be computed for each subgroup, capturing differential borrowing and informing the degree of shrinkage (Neuenschwander et al., 2019).

In decision analysis, particularly value-of-information calculations using EVSI, direct estimation of prior ESS enables tractable and accurate estimation of informational value under complex uncertainty and dependence structures (Li et al., 2024).

7. Practical Considerations and Recommendations

  • Avoid arbitrary or “conventional’’ sets for ESS parameters in Bayesian network learning. Instead, perform a sensitivity analysis over plausible ranges or apply model selection protocols integrating over ESS (Silander et al., 2012).
  • For nonconjugate Bayesian analyses, use the ELIR definition for prior and posterior ESS due to its unique property of predictive consistency.
  • When estimating ESS computationally, nonparametric regression-based methods offer state-of-the-art accuracy at high computational efficiency, provided their assumptions are met (Li et al., 2024).
  • In trial designs, explicitly account for ESS when combining external and internal data, as this directly impacts required follow-up, power, and interpretation (Neuenschwander et al., 2019).
  • For vector-valued parameters or hierarchical models, compute ESS for each marginal or adapt multivariate information-theoretic approaches; joint ESS definitions remain an open research topic (Li et al., 2024).

Equivalent sample size thus serves as a unifying currency for prior and posterior information, central to both Bayesian inference and principled experimental design. Its practical use requires rigorous definition, careful estimation, and nuanced interpretation in context-specific modeling frameworks.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Equivalent Sample Size.