Open-faced Sandwich Adjustment
- Open-faced Sandwich Adjustment is a post-hoc transformation that corrects the covariance of MCMC draws when using surrogate objective functions.
- It employs sandwich asymptotics to scale and rotate quasi-posterior samples, aligning them with the true asymptotic variance.
- The method is applicable in spatial statistics, robust modeling, and complex systems, offering valid frequentist confidence intervals despite non-likelihood data.
The open-faced sandwich (OFS) adjustment is a post-hoc transformation designed for Markov chain Monte Carlo (MCMC) methods when the traditional likelihood function is unavailable, intractable, or undesirable. In scenarios where only surrogate or quasi-likelihoods—objective functions maximized for asymptotically normal estimators—are accessible, standard Bayesian MCMC sampling yields posteriors with the incorrect curvature, resulting in poor frequentist uncertainty quantification. The OFS adjustment uses sandwich asymptotics to rotate and scale the raw MCMC draws such that their empirical distribution matches the correct asymptotic variance, enabling valid confidence sets even with non-likelihood objective functions (Shaby, 2012).
1. Motivation and Context
Frequently, applications in spatial statistics, robust modeling, or extreme value theory encounter intractable, unstable, or unknown likelihoods. Standard alternatives, such as composite or tapered likelihoods, or other estimating-equation criteria, produce consistent and asymptotically normal point estimators, denoted . Directly substituting these objectives into the MCMC acceptance ratio creates a "quasi-posterior," but this fails to yield draws with the correct spread.
Specifically, when using a non-likelihood surrogate in place of the log-likelihood , the resulting quasi-posterior may concentrate on the correct value, but its covariance misrepresents the estimator's sampling variability. This leads to credible intervals with poor frequentist coverage, undermining Bayesian inferential procedures when the true likelihood is inaccessible.
2. Theoretical Foundation: Extremum Estimators and Sandwich Variance
Consider an objective function formed from observations, with extremum estimator . Under regularity,
where the sandwich information matrices are
- ,
- ,
- so the sandwich variance .
For quasi-posterior outputs, the curvature is governed by , not . Naïve use of the quasi-posterior yields intervals with the wrong width unless this discrepancy is corrected.
3. Construction and Application of the OFS Adjustment
The OFS adjustment is formalized as a post-hoc linear transformation aligning the quasi-posterior draws’ covariance with the correct sandwich variance.
- Adjustment Matrix:
The goal is to find a matrix such that if , then . This is satisfied by setting
where and denote matrix square roots. In practical applications, estimated versions , are substituted to form .
- Adjusting MCMC Draws:
- Generate MCMC samples from the quasi-posterior .
- Compute the point estimate (e.g., mean or mode).
- Apply the OFS transformation to each draw:
- The empirical quantiles of provide asymptotically valid confidence intervals.
- Integration in Gibbs Samplers:
When is present in only some conditional blocks, block-specific adjustments can be computed and embedded within the corresponding Gibbs updates. A recommended two-pass approach runs the unadjusted sampler to estimate , then repeats with OFS correction, sufficing in most contexts.
4. Asymptotic Properties and Validity
The OFS adjustment is theoretically justified under regularity conditions ensuring the consistency and asymptotic normality of both the extremum estimator and the quasi-posterior [Chernozhukov & Hong 2003]. The result is that the OFS-adjusted draws share the same first-order asymptotic distribution as the extremum estimator , yielding quantiles with correct frequentist coverage for any fixed . Necessary conditions include the smoothness of , identifiability, invertibility of and , and increasing sample size.
5. Estimation of Information Matrices and
Accurate computation of the OFS adjustment requires estimates of and :
- Estimating :
- Option 1: Finite-difference approximation of the Hessian at , using stored values of .
- Option 2: Use the sample covariance of the MCMC draws as an estimate for .
- Option 3: Insert into an analytic formula for , if available.
- Estimating :
- For i.i.d. data , the empirical variance of gradients is
- If closed-form is obtainable, plug in . - Otherwise, use a parametric bootstrap: simulate from model, compute gradients, and aggregate to estimate variance.
- Computational Considerations:
Hessian and bootstrap calculations introduce overhead, but only need to be computed once per chain (or per block in a Gibbs sampler). Proper tuning (for example, avoiding instability in finite-differenced Hessians) and possible regularization are advised, especially when or are ill-conditioned.
6. Application Examples
OFS adjustment has demonstrated empirical efficacy in large-scale and complex models:
| Example | Model/Objective Function | Result |
|---|---|---|
| Covariance tapering for Gaussian processes | Tapered log-likelihood (sparse Σ⊙T) | OFS achieves near-nominal coverage; naïve intervals severely miscover |
| Poisson spatio-temporal GLMM for eBird data | Tapered likelihood + Gibbs sampling | OFS corrects under/over-dispersion; credible bands match ecological signals |
In spatial Gaussian process applications, direct evaluation of the likelihood incurs computational cost; replacing this with a tapered log-likelihood allows for scalable estimation. Analytical expressions for and facilitate OFS computation, with adjusted intervals providing frequentist-valid inference. In spatio-temporal generalized linear mixed models for citizen science counting data (eBird), blockwise OFS correction implemented inside a Gibbs sampler propagates appropriate uncertainty, improving quantile coverage and inference quality (Shaby, 2012).
7. Implications and Summary
The open-faced sandwich adjustment provides a rigorous framework for extending MCMC-based Bayesian procedures to models where only non-likelihood objective functions are available. It enables practitioners to retain hierarchical modeling flexibility and uncertainty propagation, while ensuring that posterior intervals have desirable frequentist properties. The method’s success hinges on accurate estimation of the sandwich matrices, and its structured post-hoc or “block-wise” adjustments are broadly applicable, including to tapered Gaussian processes, composite likelihoods, and spatio-temporal GLMMs (Shaby, 2012). A plausible implication is that as computational and modeling challenges drive increased use of quasi-likelihoods, OFS-style corrections will become essential for valid uncertainty quantification in Bayesian-like workflows.