Bayesian Spike-and-Slab Framework
- Bayesian Spike-and-Slab Framework is a hierarchical model that combines a sharp 'spike' at zero with a diffuse 'slab' to represent coefficients for sparse estimation.
- It enables effective variable selection and uncertainty quantification in high-dimensional regression by inducing a multimodal posterior over sparse supports.
- The framework leverages restricted isometry properties and efficient rejection sampling techniques to ensure accurate, scalable posterior sampling with formal computational guarantees.
A Bayesian Spike-and-Slab Framework is a hierarchical probabilistic modeling approach for sparse estimation, variable selection, and uncertainty quantification in high-dimensional inference problems. The framework combines a "spike" component (typically a point mass at zero or a sharply peaked continuous density) with a "slab" component (a diffuse or heavy-tailed density) as a prior distribution over model coefficients, supports, or structural parameters. This yields a multimodal posterior that encodes combinatorial uncertainty over sparsity patterns and continuous uncertainty over effect sizes, with formal guarantees and efficient computational algorithms now available for exact posterior sampling in regimes previously inaccessible to traditional methods (Kumar et al., 4 Mar 2025).
1. Model Specification and Prior Construction
Bayesian spike-and-slab regression is defined via observations , with , , and an unknown sparse . The canonical spike-and-slab prior is
where each coordinate is zero ("spike") with probability or drawn i.i.d. from the slab density (e.g., or Laplace) with probability (Kumar et al., 4 Mar 2025). This induces a prior over supports : The prior is thus both discrete (over possible supports) and continuous (over effect sizes in the slab).
2. Posterior Characterization and Analytic Structure
Under Gaussian noise and a Gaussian slab, the posterior takes the form: In the Gaussian-slab case, one obtains an explicit mixture representation: with explicit formulas for the mixture means, covariances, and weights (Kumar et al., 4 Mar 2025):
For Laplace slab, closed-form Gaussian integrals are lost, but the same underlying mixture structure persists, subject to numerical integration (Kumar et al., 4 Mar 2025).
3. Sampling-Complexity, Restricted Isometry and Statistical Regimes
A key insight is that posterior sampling, in order to be accurate and tractable in high dimensions, demands that satisfy a restricted isometry property (RIP) up to sparsity , where is the expected sparsity and is total variation error tolerance (Kumar et al., 4 Mar 2025). For Gaussian ,
- Polynomial-time, high-accuracy sampler: achieves TV error in operations.
- Near-linear-time sampler: achieves the same accuracy in (Kumar et al., 4 Mar 2025).
RIP suffices for standard random matrix ensembles (sub-Gaussian, subsampled Fourier, etc.), requiring only a sublinear scaling of samples in the dimension . These bounds break past barriers of strictly linear sample regimes or strong SNR assumptions common in previous literature.
4. Algorithmic Advances: Hint-Vector Estimation and Product Proposals
The sampling framework proceeds in two stages:
- Hint-vector estimation: Fast sparse-recovery ( or -based) yields an initial estimate with support . With high posterior probability, for sample support and small (Proposition 2.8).
- Coordinatewise product-proposal + rejection sampling: Using "recentering" (Lemma 3.3), weights over supports become simple coordinatewise products. A conditional Poisson product distribution over can be sampled in time and is provably close (within constant ratios, Lemma 3.7) to the true posterior mass, allowing TV-accurate rejection sampling to the true posterior support (Kumar et al., 4 Mar 2025).
Once support is drawn, sampling is exact (Lemma 3.11).
5. Provable Posterior Guarantees, Sparsity, and Estimation-to-Sampling Results
The main theorem (Kumar et al., 4 Mar 2025): Under stipulated RIP and sample size,
The computational cost is as above, and rejection sampling mixes efficiently over the posterior support (mixing cost ). The near-linear-time sampler requires samples and achieves comparable accuracy.
Structural lemmas include:
- Support-sparsity (Corollary 2.10): For any product prior, .
- Posterior-to-estimation (Proposition 2.8): Any estimator that is good in metric with high probability induces a sampling procedure that, with probability , draws -samples within double the estimation error in .
6. Extension to Laplace Slabs
For slab , the Gaussian mixture structure lacks closed-form integrals. The algorithm adapts by:
- Using Monte Carlo or annealing-based normalizer estimation for each mixture component (Prop 4.1, Cor 4.6) to accuracy in .
- Restricting to to control Laplace tail errors (Lemma 4.2).
Theorem 1.3 (Cor 4.14): For , , , sample in time with TV. Near-linear-time algorithms analogous to the Gaussian case hold for (Kumar et al., 4 Mar 2025).
7. Technical and Applied Impact
This framework supplies the first polynomial-time, sublinear-measurement, provably exact samplers for spike-and-slab posteriors in high-dimensional sparse linear regression, valid for all SNRs, with flexible extension to Laplace diffuse priors (Kumar et al., 4 Mar 2025). The approach unifies continuous algorithms, precise RIP-based concentration, and rejection-sampling with conditional-Poisson tractable supports for rigorous total variation-accuracy and running-time guarantees.
The framework establishes spike-and-slab posterior sampling—with certifiable accuracy and scalable computation—as the theoretical and practical gold standard for Bayesian sparse regression in high dimensions.