Multi-Stage Recursive Bayesian Approach
- Multi-stage recursive Bayesian approaches are methods that bridge the prior and posterior using a sequence of intermediate distributions to enhance evidence estimation.
- They employ sequential updates, importance sampling, and thermodynamic integration to reduce estimator variance and enable efficient prior sensitivity analysis.
- These techniques are applied in high-dimensional problems such as mixture modeling and astronomical analysis, demonstrating practical benefits in computational Bayesian model selection.
A multi-stage recursive Bayesian approach is a class of Bayesian computational strategies in which inference for a complex model, or estimation of model evidence, is performed via a sequence of updates across multiple intermediate distributions, stages, or data partitions. Rather than conducting inference in a single monolithic step, the approach recursively builds up the target posterior or normalizing constant by leveraging a chain of conditional decompositions, tempering schemes, or scenario splits. This not only enhances computational efficiency but also enables new forms of sensitivity analysis and model diagnostics, especially regarding prior or stagewise choices.
1. Foundations: Recursive Pathways and Bridging Sequences
The central construct in multi-stage recursive Bayesian approaches is the interpolation between the prior and the posterior by means of a sequence of intermediate (often tempered) distributions. These are indexed as , where is the unknown normalizing constant and encodes the unnormalized density for stage . Notably:
- (with ),
- (the unnormalized posterior with ).
A canonical bridging sequence is the family of power posteriors, where
Recursive estimators such as “biased sampling,” “reverse logistic regression” (RLR), and the “density of states” method proceed by simulating samples from each , pooling these into a grand sample, and solving a set of recursive updates to estimate the 's.
The recursive update for the normalizing constants is (for ):
where and is the th draw from the th distribution (Cameron et al., 2013). This estimator arises in various equivalent forms, such as via maximization of a quasi–log–likelihood in RLR, or as an importance sampling identity based on a pseudo-mixture density.
2. Connection to Marginal Likelihood Estimation and Prior Sensitivity
By expressing the overall marginal likelihood (evidence) as a telescoping product across the bridge (), the recursive pathway efficiently estimates the model evidence in a high-dimensional space. A crucial computational benefit is that, after the bridging sample is computed, the same particles can be reused for efficient prior-sensitivity analysis.
Given the pooled pseudo-mixture density
importance sampling reweighting allows marginal likelihoods for alternative priors to be computed without rerunning the sampling, via
This enables rapid assessment of evidence and Bayes factors under varying modeling assumptions—a valuable property in Bayesian model selection contexts subject to prior sensitivity.
3. Thermodynamic Integration via Importance Sampling (TIVIS) Perspective
A notable theoretical contribution is the reinterpretation of the recursive estimator as a form of thermodynamic integration via importance sampling (TIVIS). Here, the log ratio between two successive normalizing constants is expressed as an integral over the “temperature” parameter :
where
Numerical integration of this quantity (e.g., via Simpson’s rule) yields an estimator equivalent to the recursive update above. The TIVIS view exposes the dependency of estimation error on divergence between successive levels, thereby informing optimal tempering schedule design.
4. Relation to Nested Sampling and Equivalent Integral Representations
The recursive pathway framework is closely linked to the nested sampling technique, which also estimates evidence by compressing the prior mass via constrained likelihood levels. In nested sampling, the multidimensional evidence integral is reduced to a one-dimensional integral over “prior mass” :
or, in terms of a change of variable to “energy” , as
where is the density at energy level . The normalization steps in biased sampling or RLR correspond to sequential contractions of prior mass in nested sampling; thus, asymptotic variance and convergence analyses for biased sampling directly inform the properties of nested sampling estimators. This equivalence is highly relevant for high-dimensional and multimodal problems, such as those encountered in cosmological or complex mixture modeling.
5. Implementation Considerations and Computational Strategies
Multi-stage recursive Bayesian approaches enable substantial computational gains by decoupling the challenging task of direct sampling from the posterior from the auxiliary task of normalizing constants computation. Simulating from a suite of intermediate distributions (e.g., power posteriors) allows broader coverage of the parameter space than can be achieved solely by sampling the prior or posterior.
After forming the bridging sample:
- The recursive estimator’s key quantities (such as and ) can be efficiently computed vectorized over the sample.
- Retrospective importance sampling for prior-sensitivity analysis is efficient and does not require additional sampling.
- The TIVIS formulation offers a diagnostic for whether the chosen bridging path will yield low- or high-variance recursive estimators.
In practice, this strategy imposes moderate requirements: a set of intermediate distributions, manageable sampling from each, and pooling of all samples. Convergence monitoring is typically achieved by examining the stability of normalizing constant estimates or the effective sample sizes under .
6. Practical Applications and Impact
Multi-stage recursive Bayesian techniques have become valuable in computational Bayesian model selection, especially where direct computation of marginal likelihood is intractable and where sensitivity to prior specification is an unresolved issue. Their ability to “recycle” bridging samples for prior-sensitivity analysis addresses a longstanding challenge in applied Bayesian workflow. The recursive framework underpins a broad array of methods in computational statistics and physics, unifying seemingly distinct algorithms with a shared normalization strategy.
Applications highlighted in the literature include:
- Mixture modeling (e.g., the galaxy dataset),
- Astronomical model selection,
- Bayesian variable selection problems,
- General model evidence computation for hierarchical and latent variable models.
The recursive pathway approach, its TIVIS extension, and the established connections to nested sampling are now foundational elements in the modern Bayesian computational toolbox, supporting efficient, robust, and interpretable evidence estimation and prior sensitivity analysis (Cameron et al., 2013).