Rejection Sampling Method
- Rejection-sampling-based methods are Monte Carlo techniques that sample from target distributions using tractable proposals and accept/reject steps to ensure correctness.
- They enable efficient sequential Bayesian updates by tracking posterior moments, drastically reducing memory requirements in high-dimensional settings.
- Extensions such as envelope clipping in approximate rejection sampling balance computational efficiency with controlled estimation error for online inference.
A rejection-sampling-based method comprises a class of Monte Carlo algorithms for drawing independent samples from a target probability distribution, typically specified up to normalization, by leveraging samples from a tractable proposal (or envelope) distribution and an accept/reject criterion based on a ratio of densities. The method is foundational in computational statistics, Bayesian inference, and scientific simulation, with theoretical guarantees on correctness and broad applicability to both discrete and continuous spaces. Its modern formulations address efficiency, adaptivity, memory savings, extensions beyond log-concave domains, and integration with particle filtering for online sequential inference.
1. Classical Rejection Sampling: Principle and Algorithm
Given a latent parameter with known prior and likelihood , the goal is to generate samples from the posterior . The standard rejection sampling algorithm proceeds as follows:
- Setup: Choose a proposal density such that for all , with minimal . In Bayesian contexts, often .
- Sampling: Repeatedly:
- Draw .
- Draw .
- Accept if ; else reject and repeat.
- Correctness: Accepted samples are exactly distributed according to the target posterior.
The acceptance probability is , where . Although exact, classical rejection sampling is intractable for sharply peaked likelihoods or high-dimensional since can grow rapidly, lowering acceptance exponentially (Wiebe et al., 2015).
2. Rejection Filtering: Moment Tracking with Particle Updates
Rejection filtering augments rejection sampling by updating only low-order posterior moments (mean , covariance ) with each batch of accept/reject trials, instead of storing all accepted samples:
- For trials, accumulate sums , , count .
- For accepted , update and ; .
- At batch end: if , estimate , . If , inflate by .
This "particle filtering" for moments delivers an memory cost—dramatically less than for storing all samples, where is parameter dimension and desired accuracy (Wiebe et al., 2015).
3. Approximate Rejection Sampling: Envelope Clipping
Exact may be impractically large or unknown. Approximate rejection sampling uses an envelope with , accepting with probability . Over-acceptance occurs for where , but total error is controlled:
If the over-accepted mass is small, the Hellinger distance between the approximate and true posterior is (Wiebe et al., 2015). Acceptance rate remains at least .
4. Complete Rejection Filtering Algorithms
Exact Rejection Filtering (RFUpdate):
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 |
Input: prior p(θ), prior moments (μ, Σ), evidence E, envelope κ_E, trials m, recovery r
Output: posterior moments (μ_new, Σ_new), N_a
M←0, S←0, N_a←0
for i = 1..m:
θ ∼ p(θ), u∼Uniform(0,1)
if u ≤ p(E|θ)/κ_E:
M←M+θ, S←S+θ θᵀ, N_a←N_a+1
if N_a>0:
μ_new←M/N_a
Σ_new←[S−N_a μ_new μ_newᵀ]/(N_a−1)
return (μ_new,Σ_new,N_a)
else:
μ_new←μ
Σ_new←(1+r)Σ
return (μ_new,Σ_new,0) |
Approximate Rejection Filtering: As above, with accept-test .
5. Computational Guarantees and Empirical Performance
Under efficient prior and uniform sampling, and bounded , the algorithm runs in time per update and bits to -precision. Memory savings over particle filters are significant in high dimensional or streaming settings.
Empirical evaluations demonstrate:
A. Frequency Tracking (active experiments):
- Unknown oscillator phase .
- With proposals per update and kbit memory, rejection filtering achieves RMSE , matching or exceeding SMC (Liu–West) with particles (100× more memory).
B. MNIST Classification (feature querying):
- Active selection of pixel intensities.
- Stopping thresholds , error rates (even/odd), outperforming kNN under identical feature budgets.
- Feature importance naturally arises from query frequencies; removing lowest-frequency pixels preserves accuracy.
Key findings: rejection filtering supports sequential Bayesian updates on memory-constrained devices, enables active learning and feature selection, and is robust even with loose envelope choices (Wiebe et al., 2015).
6. Advantages, Limitations, and Domains of Application
Advantages:
- Memory efficiency: O(), orders of magnitude superior for large .
- Asymptotic correctness: Maintains correct posterior summaries under exact envelope.
- Robustness to approximate envelopes: Controlled error when envelope bounds are not tight.
- Online and active settings: Applicable to time-dependent, data-adaptive Bayesian inference.
Limitations:
- Very small acceptance probabilities in high dimensions or extremely peaked likelihoods may still pose practical barriers.
- Parametric approximation: Assumes that posterior is well captured by first/second moments, e.g., Gaussian form; multimodal or strongly non-Gaussian posteriors may require other strategies.
Domain of Application:
- Sequential Bayesian parameter estimation, streaming inference, online classification, and settings where both computational and storage costs are constraints.
7. Summary of the Rejection-Sampling-Based Method Innovations
Rejection filtering generalizes classical rejection sampling by incrementally updating estimates of posterior moments via efficient trial batches, not cloud storage of samples. With both exact and envelope-clipped variants, it dramatically reduces required memory, preserves correctness, and is robust to envelope misspecification. Its empirical performance matches or exceeds standard particle filters in time-dependent and feature-selection scenarios, enabling efficient Bayesian learning under compute/memory constraints and online inference (Wiebe et al., 2015).