Bayesian Sample Influence Measures
- Sample influence measures are quantitative diagnostics in Bayesian analysis that assess the sensitivity of posterior results to changes in individual observations via likelihood and prior perturbations.
- They leverage MCMC simulation to compute variance of log-likelihoods and replicate predictive distributions, enabling effective detection of influential and outlying data points.
- The approach supports model validation and refinement by quantifying prior-data conflict and facilitating robust outlier identification in both classical and high-dimensional models.
Sample influence measures are quantitative diagnostics designed to assess the sensitivity of Bayesian inferential results to individual observations or subsets of the data. In the Bayesian context, influence and leverage measure how perturbations to likelihood or prior components associated with observations affect the posterior or derived predictive quantities. By formalizing and computationally realizing these ideas for general Bayesian models—including those fitted with Markov chain Monte Carlo (MCMC)—researchers are equipped to diagnose influential or outlying data points, assess model robustness, and interpret predictive information criteria. These diagnostics also support detection of prior-data conflict, supplementing both model validation and selection procedures.
1. Bayesian Leverage and Influence: Core Definitions
The Bayesian local influence for an observation is defined through the effect of a small change in the data-point’s weight in the likelihood. Letting the observation weights be in a pseudo-likelihood
the Kullback–Leibler divergence between the posterior () and its perturbed version quantifies observation-level sensitivity. Taylor expansion yields a quadratic approximation: where under the posterior. The Bayesian local influence of observation is thus
which expresses posterior uncertainty in the individual log-likelihood.
Bayesian leverage is defined in relation to replication at the design point . The leverage (also called the Bayesian hat-value) is
quantifying how much observing an additional at would alter predictive expectations. For the linear model, reduces to the frequentist hat value.
Practical deletion or upweighting influence diagnostics are constructed by finite perturbations: the “doubling influence” (weight set to 2), and the “zeroing influence” (weight set to zero).
2. MCMC Computation of Diagnostics
The computational implementation leverages standard MCMC simulation. Given draws from :
- The variance–covariance matrix is estimated as the sample covariance of across .
- Leverage is estimated via a two-sample Monte Carlo, by drawing pairs from independent MCMC runs and computing the mean KL divergence between the predictive distributions for the replicate .
For multivariate perturbations (subsets), the same machinery applies, allowing construction of matrices for joint or conditional influence and subsequent eigenanalysis to detect clusters of jointly influential or maskingly anomalous observations.
These diagnostics are fast, requiring only the MCMC samples obtained in the main analysis, and the necessary computations are simple statistics over the log-likelihood trace.
3. Outlier Detection: Leverage, Influence, and Conformal Measures
A key insight is that a distinction between leverage and influence facilitates robust outlier identification. High leverage () indicates that is unusual in predictor space, irrespective of , while high influence () corresponds to an outcome exerting abnormal effect on the posterior.
The conformal local outlier statistic (Editor’s term) is introduced: A high identifies points that are more influential than their leverage would suggest and is thus suited to general-purpose outlier detection regardless of whether they substantially affect the overall fit.
4. Relationship to Predictive Information Criteria
Local influence and leverage diagnostics are shown to be directly related to model complexity penalties in standard Bayesian predictive criteria.
- The WAIC optimism penalty is
making local influence the additive component of model “optimism.”
- The DIC effective number of parameters can be represented by the sum of leverages:
- An alternative penalty, , provides further information about the “global” variability inherent in the data-likelihood.
5. Prior–Data Conflict Diagnostic and Cross-Conflict Assessment
The ratio is introduced as a diagnostic for systematic conflict between the prior and the observed data. In the presence of a diffuse prior, . However, when the prior pulls the posterior mean away from the MLE, increases. Substantially flags prior-data tension or cross-conflict between data subsets in hierarchical models, motivating further posterior examination or modularization.
This diagnostic supplements traditional model checks and is applicable to complex Bayesian models including those with hierarchical or structured priors.
6. Practical Implications and Applications
Bayesian influence and leverage diagnostics support data-quality assessment, model checking, and robust summary inference. Leveraging the diagnostic matrices (influence) and (leverage), one can:
- Identify and summarize individual and subgroup anomalies (“masking” and “swamping” are resolved via joint and conditional perturbation matrices).
- Guide model refinement and data cleaning, detecting problematic cases that may not dramatically alter posterior means but signal local fit issues.
- Interpret complexity penalties in predictive criteria in terms of sample-wise and population-level contributions.
Eigenanalysis of these matrices extends outlier detection to multivariate and high-dimensional settings, enabling detection of “collective” aberrance.
7. Illustrative Examples and Computational Considerations
Empirical demonstrations on both classical benchmark datasets (Abalone, Bike Sharing) and substantive applications (e.g., survival analysis with UNOS data) confirm the diagnostics' utility in practice. Calculations require only standard posterior draws and basic vector operations, ensuring scalability to moderately high-dimensional Bayesian models commonly encountered in applied research.
In summary, Bayesian sample influence measures constitute a rigorous and computationally tractable framework for model sensitivity analysis, integrating seamlessly with posterior simulation and predictive modeling workflows (Plummer, 25 Mar 2025).