Posterior contraction in sparse Bayesian factor models for massive covariance matrices
Published 16 Jun 2012 in math.ST | (1206.3627v4)
Abstract: Sparse Bayesian factor models are routinely implemented for parsimonious dependence modeling and dimensionality reduction in high-dimensional applications. We provide theoretical understanding of such Bayesian procedures in terms of posterior convergence rates in inferring high-dimensional covariance matrices where the dimension can be larger than the sample size. Under relevant sparsity assumptions on the true covariance matrix, we show that commonly-used point mass mixture priors on the factor loadings lead to consistent estimation in the operator norm even when $p\gg n$. One of our major contributions is to develop a new class of continuous shrinkage priors and provide insights into their concentration around sparse vectors. Using such priors for the factor loadings, we obtain similar rate of convergence as obtained with point mass mixture priors. To obtain the convergence rates, we construct test functions to separate points in the space of high-dimensional covariance matrices using insights from random matrix theory; the tools developed may be of independent interest. We also derive minimax rates and show that the Bayesian posterior rates of convergence coincide with the minimax rates upto a $\sqrt{\log n}$ term.
The paper derives precise posterior contraction rates for estimating high-dimensional covariance matrices using sparse Bayesian factor models, achieving near minimax optimality.
It compares point mass mixture and continuous shrinkage priors, revealing that continuous shrinkage provides practical and robust performance in ultra-high-dimensional settings.
The study utilizes novel test constructions based on random matrix theory and simulation experiments to validate its theoretical guarantees and computational effectiveness.
Posterior Contraction in Sparse Bayesian Factor Models for Massive Covariance Matrices
Introduction and Problem Setting
This work addresses the theoretical properties of posterior contraction in sparse Bayesian factor models for estimating high-dimensional covariance matrices, specifically focusing on situations where the ambient dimension p significantly exceeds sample size n (p≫n). The focus is on latent factor models for covariance estimation under sparsity: each observed vector yi∈Rp is modeled as
yi=Ληi+εi,εi∼Np(0,Ω),
where Λ is a p×k factor loading matrix (with k≪p), ηi∼Nk(0,Ik), and Ω is diagonal. Marginally, the covariance takes the reduced form Σ=ΛΛT+Ω, drastically reducing the number of free parameters from O(p2) to O(pk). The key practical and theoretical challenge is to obtain non-trivial estimation guarantees when p grows much faster than n.
The analysis emphasizes "ultra"-high-dimensional settings and provides precise posterior contraction rates under sparsity assumptions and prior constructions that are relevant for genomic, neuroscience, and other modern high-dimensional datasets.
Assumptions and Prior Structures
The theoretical development rests on several assumptions:
Factor Structure on Truth: The true covariance has the form Σ0n=Λ0nΛ0nT+Ω0n, with growing pn and possibly growing k0n.
Column Sparsity: Each column of Λ0n has at most sn nonzero components (sn≪pn), reflecting factor sparsity.
Pervasiveness and Conditioning: Spectral conditions on Λ0n, echoing "pervasive" factors in random matrix theory, with mild restrictions on the growth of the largest eigenvalue cn and minimal conditions on the residual variance.
Sample Size and Model Complexity: Conditions such that cnk0n3/2(snlogpn/n)logn→0, which allow pn to be of order exp(nα) for some α∈(0,1/5) under typical sparsity regimes.
Two types of priors for loadings are analyzed:
Point Mass Mixture Priors (Spike-and-Slab): Each entry is zero with high probability, otherwise drawn from a heavy-tailed distribution. This matches frequentist penalization with exact sparsity but is computationally challenging.
Continuous Shrinkage Priors: Hierarchical scale mixtures of Laplace (double exponential) distributions with global-local scales (motivated by the Horseshoe, Dirichlet-Laplace, etc.). These enable efficient MCMC without exact zeros.
Priors on k favor small numbers of factors, and on σ2 (residual variance) are diffuse but proper.
Main Theoretical Results
Posterior Contraction Rates
A central contribution is the derivation of explicit posterior contraction rates for estimating the covariance Σ0n in the operator (spectral) norm, under both point mass mixture and continuous shrinkage priors. The results can be summarized as follows:
Operator Norm Consistency: If sn≳logpn and k0n=O(1), the posterior contracts at rate
εn=cnnsnlogpnlogn
in operator norm, i.e.,
Πn(∥Σn−Σ0n∥2>Mεn∣y(n))→0
in probability, for any sufficiently large M. The dependence on cn encodes the "energy" in the largest eigenvalue.
General k0n Growth: If k0n is allowed to grow, the contraction rate becomes
cnk0n3/2nsnlogpnlogn.
Thus, sparsity and factor proliferation both contribute to statistical complexity.
Matching Minimax Lower Bounds: For fixed k0n, the posterior rates coincide with a new minimax lower bound (proved via Fano's method) up to an explicit logn factor:
Σ^ninfΣ0nsupE∥Σ^n−Σ0n∥2≥cnsnlogpn/n.
This demonstrates the theoretical efficiency of the Bayesian procedures under the specified conditions.
Robustness to Prior Specification: The optimal contraction rates are obtained for both point mass mixture and the proposed shrinkage priors, provided the priors allocate sufficient mass to neighborhoods of the true sparse loading vectors.
Priors and High-Dimensional Properties
The paper develops properties of the analyzed continuous shrinkage priors that guarantee:
Prior Concentration: Lower bounds for the prior probability of small balls around arbitrary s-sparse vectors comparable to those for point mass mixture priors.
Effective Dimensionality Control: Exponential decay of the probability that the number of "large" (> δ) entries in the loading is much bigger than order sn.
Tail Control: Subexponential deviation bounds on the ℓ1 norm of the priors, ensuring concentration on "reasonable" model sizes.
These results are nontrivial as traditional global-local shrinkage mechanisms do not guarantee the necessary localized prior mass in ultra-high dimensions.
Test Construction and Proof Techniques
The authors advance the theory of posterior contraction in non-Hellinger metrics by constructing nonparametric tests for covariance matrices using techniques from random matrix concentration inequalities. The construction leverages the fact that, under the factor model, the dominant contribution to the operator norm comes from the low-rank term, enabling projection-based tests with exponentially decaying type I and II errors in high dimensions.
Proofs build on detailed metric entropy calculations, sharp bounds on prior masses, and control of test errors, encompassing random matrix theory and empirical process theory tools.
Numerical Experiments
The simulation studies compare the proposed Bayesian methods with frequentist techniques such as POET and adaptive thresholding, evaluating covariance estimation accuracy in operator norm under varying p,n,k0n,sn, for both well-specified and misspecified noise structures.
The continuous shrinkage prior outperforms or matches other methods in all settings, especially as model complexity increases, and is robust to the absence of exact sparsity or diagonal noise. Point mass mixture priors deteriorate in very large models, primarily due to computational mixing issues. The empirical results thus reinforce the theoretical findings.
Implications and Future Directions
The theoretical advances provide a framework for principled Bayesian inference for covariance estimation under realistic high-dimensional settings with latent structure and sparsity. The results justify the use of continuous shrinkage priors as computational surrogates for spike-and-slab approaches, with guarantees matching minimax frequentist rates up to log-factors.
Several avenues for further research are apparent:
Extension to Approximate Factor Models: Relaxing structural assumptions to allow for non-diagonal idiosyncratic variance and weakly sparse low-rank structure.
Adaptive or Empirical Bayes Procedures: Prior hyperparameter tuning for optimal adaptation to unknown sparsity and eigenvalue growth.
Posterior Convergence for Functional Parameters: Extension to linear or quadratic functionals of high-dimensional covariance (e.g., prediction under factor models).
Sharper Minimax Adaptivity: Investigation of potential improvements to eliminate the logn factor and extensions to more general sparsity regimes.
Conclusion
This work provides a rigorous analysis of posterior contraction properties for a broad class of Bayesian factor models under sparsity, for covariance estimation in "ultra"-high-dimensional settings. By establishing precise rates, matching minimax lower bounds, and demonstrating practical computational robustness, it strongly supports the use of Bayesian shrinkage methods, both theoretically and empirically, in modern high-dimensional statistical inference (1206.3627).