Pseudo-Marginal Metropolis–Hastings (PMMH)
- PMMH is a Markov chain Monte Carlo method that uses unbiased estimators to perform exact Bayesian inference on models with intractable likelihoods.
- It employs techniques such as signed, block–Poisson, and correlated estimators to efficiently handle doubly intractable problems and reduce variance.
- Practical implementations of PMMH require careful tuning of block sizes and estimator parameters, as demonstrated in applications like Ising and Kent models.
A pseudo-marginal Metropolis–Hastings (PMMH) algorithm is a Markov chain Monte Carlo (MCMC) method for exact inference in models with intractable likelihoods, where only unbiased non-negative estimators of the unnormalized target density are available. PMMH encompasses a class of algorithms, including standard, correlated, and signed estimators for settings such as doubly intractable distributions. These algorithms have been extensively studied for both theory and high-dimensional applications, notably in latent variable models, doubly intractable posteriors, and @@@@1@@@@ (Yang et al., 2022).
1. General Structure and Exactness
Let denote the observed data, the parameter of interest, and its prior. If is intractable but one can compute an unbiased estimator using an auxiliary random vector with , define the extended target as
The PMMH transition accepts with probability
The marginal chain in leaves the exact posterior invariant due to unbiasedness and standard MH theory (Yang et al., 2022).
2. Signed and Block–Poisson Estimators for Doubly Intractable Models
For posteriors where the likelihood includes an intractable normalizing constant , the doubly intractable posterior is
A signed unbiased estimator for expressions such as is constructed using the block–Poisson (BP) scheme. Introduce and partition the estimator: with and each an unbiased estimator of . Since can be negative, PMMH targets , collecting sign at each iteration and using the importance-sampling corrected estimator: as (Yang et al., 2022).
3. Correlated PMMH and Blockwise Updates
Variance in the log-likelihood ratio degrades PMMH mixing as dimension/amount of data increases, motivating correlated PMMH approaches (Deligiannidis et al., 2015, Yang et al., 2022, Dahlin et al., 2015). By partitioning into blocks and only refreshing one block per iteration, the correlation between log estimators at current and proposed states becomes
resulting in much lower variance of the acceptance ratio: where is a simple average of blockwise -estimates. The bias–correction terms guarantee unbiasedness for even in correlated block updates (Yang et al., 2022).
4. Tuning and Hyperparameter Selection
Optimal performance relies on tuning the BP estimator's block count , block size , and per-block average using a computational time criterion: with inefficiency factor depending on variance and (probability the estimate is positive). Closed-form expressions for and are available under normality assumptions. Practical guidelines: –100 for –0.99, , with –0.005 such that and (Yang et al., 2022).
5. Empirical Performance and Model Classes
Empirical results for Ising models and Kent distribution models with intractable normalizing constants show that correlated PMMH with block–Poisson estimators offers a substantial increase in effective sample size per second compared to non-blocked or Russian roulette–style signed estimators; for the Ising model, correlated BP PMMH achieved ESS/sec of 0.5 versus 0.2 for Russian roulette (a twofold efficiency gain). For small or “nearly singular” parameter settings (e.g., small sample size, near-symmetry in the Kent model), Bayesian PMMH outperforms method-of-moments or saddlepoint-based MLE estimators in root mean squared error (Yang et al., 2022).
6. Theoretical Guarantees and Limitations
PMMH remains asymptotically exact for any doubly intractable model with an unbiased -estimator; no perfect sampling is required. The block–Poisson structure admits high correlation in likelihood ratios and admits closed-form rules for tuning. The main limitation is computational cost: each MH step requires multiple -estimates (via AIS or high-dimensional integrals). The presence of negative estimates in the signed setting increases variance, so tuning for high is essential. Current open challenges include more efficient unbiased -estimators, adaptive block partitioning, and extension to streaming/big data contexts (Yang et al., 2022).
7. Practical Recommendations and Impact
- PMMH with block–Poisson and correlated block updates offers simulation-consistent inference for doubly intractable models, outperforming alternative methods in simulation efficiency under a broad range of challenging scenarios.
- Practical adoption requires careful tuning to ensure stable (signed) estimators and to control the computational cost per iteration.
- Future work will need to address estimator design for more complex intractable models and develop further adaptive and scalable extensions for real-world data at massive scale (Yang et al., 2022).