Dirichlet process mixtures of block $g$ priors for model selection and prediction in linear models (2411.00471v1)
Abstract: This paper introduces Dirichlet process mixtures of block $g$ priors for model selection and prediction in linear models. These priors are extensions of traditional mixtures of $g$ priors that allow for differential shrinkage for various (data-selected) blocks of parameters while fully accounting for the predictors' correlation structure, providing a bridge between the literatures on model selection and continuous shrinkage priors. We show that Dirichlet process mixtures of block $g$ priors are consistent in various senses and, in particular, that they avoid the conditional Lindley ``paradox'' highlighted by Som et al.(2016). Further, we develop a Markov chain Monte Carlo algorithm for posterior inference that requires only minimal ad-hoc tuning. Finally, we investigate the empirical performance of the prior in various real and simulated datasets. In the presence of a small number of very large effects, Dirichlet process mixtures of block $g$ priors lead to higher power for detecting smaller but significant effects without only a minimal increase in the number of false discoveries.
- Bayesian robustness modelling of location and scale parameters. Scandinavian Journal of Statistics 38, 691–711.
- Antoniak, C. E. (1974). Mixtures of dirichlet processes with applications to bayesian nonparametric problems. The annals of statistics pp. 1152–1174.
- On the beta prime prior for scale parameters in high-dimensional Bayesian regression models. arXiv preprint arXiv:1807.06539 .
- Criteria for Bayesian model choice with application to variable selection. The Annals of Statistics 40, 1550–1577.
- The formal definition of reference priors. Annals of Statistics 37, 905–938.
- The intrinsic Bayes factor for linear models. In Bayesian Statistics 5, Eds. A. P. D. J. M. Bernardo, J. O. Berger & A. F. M. Smith, pp. 25–44. Oxford Univ. Press.
- Bayes factors and marginal distributions in invariant situations. Sankhyā: The Indian Journal of Statistics, Series A pp. 307–321.
- Bertoin, J. (2006). Random fragmentation and coagulation processes, volume 102. Cambridge University Press.
- The horseshoe+ estimator of ultra-sparse signals. Bayesian Analysis 12, 1105–1131.
- Dirichlet–Laplace priors for optimal shrinkage. Journal of the American Statistical Association 110, 1479–1490.
- Ferguson distributions via pólya urn schemes. The annals of statistics 1, 353–355.
- Group inverse-gamma gamma shrinkage for sparse linear models with block-correlated regressors. Bayesian Analysis 1, 1–30.
- Hyper-g𝑔gitalic_g priors for generalized linear models. Bayesian Analysis 6, 387–410.
- Estimating optimal transformations for multiple regression and correlation. Journal of the American statistical Association 80, 580–598.
- Inference with normal-gamma prior distributions in regression problems. Bayesian Analysis 5, 171–188.
- The horseshoe estimator for sparse signals. Biometrika 97, 465–480.
- Objective Bayesian model selection in Gaussian graphical models. Biometrika 96, 497–512.
- Objective bayesian variable selection. Journal of the American Statistical Association 101, 157–167.
- Prior distributions for objective Bayesian analysis. Bayesian Analysis 13, 627–679.
- Consistent fractional Bayes factor for nested normal linear models. Journal of statistical planning and inference 97, 305–321.
- A horseshoe mixture model for bayesian screening with an application to light sheet fluorescence microscopy in brain imaging. The Annals of Applied Statistics 17, 2639–2658.
- Ferguson, T. S. (1973). A bayesian analysis of some nonparametric problems. The annals of statistics pp. 209–230.
- Robust Bayesian graphical modeling using Dirichlet t-distributions. Bayesian Analysis 9, 521–550.
- Methods and tools for Bayesian variable selection and model averaging in normal linear regression. International Statistical Review 86, 237–258.
- Power-expected-posterior priors for variable selection in Gaussian linear models. Bayesian Analysis 10, 75–107.
- Strictly proper scoring rules, prediction, and estimation. Journal of the American Statistical Association 102, 359–378.
- Gordy, M. B. (1998). A generalization of generalized Beta distributions. Technical report, Division of Research and Statistics, Division of Monetary Affairs, Federal Reserve.
- Green, P. J. (1995). Reversible jump Markov chain Monte Carlo computation and Bayesian model determination. Biometrika 82, 711–732.
- Alternative prior distributions for variable selection with very many more variables than observations. University of Kent Technical Report .
- Hans, C. (2009). Bayesian lasso regression. Biometrika 96, 835–845.
- Adaptive lasso for sparse high-dimensional regression models. Statistica Sinica pp. 1603–1618.
- On the use of non-local prior densities in Bayesian hypothesis tests. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 72, 143–170.
- Bayesian model selection in high-dimensional settings. Journal of the American Statistical Association 107, 649–660.
- A reference Bayesian test for nested hypotheses and its relationship to the Schwarz criterion. Journal of the American Statistical Association 90, 928–934.
- Continuous shrinkage prior revisited: a collapsing behavior and remedy. arXiv preprint arXiv:2007.02192 .
- Bayesian adaptive lasso. Annals of the Institute of Statistical Mathematics 66, 221–244.
- Variable selection using shrinkage priors. Computational Statistics & Data Analysis 107, 107–119.
- Mixtures of g-priors in generalized linear models. Journal of the American Statistical Association 113, 1828–1845.
- Mixtures of g-priors for Bayesian variable selection. Journal of the American Statistical Association 103, 410–423.
- Rejection sampling for an extended gamma distribution. Unpublished manuscript .
- Neal, R. M. (2000). Markov chain sampling methods for dirichlet process mixture models. Journal of computational and graphical statistics 9, 249–265.
- O’Hagan, A. (1995). Fractional Bayes factors for model comparison. Journal of the Royal Statistical Society: Series B (Methodological) 57, 99–118.
- The Bayesian lasso. Journal of the American Statistical Association 103, 681–686.
- Local shrinkage rules, lévy processes and regularized regression. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 74, 287–311.
- Bayesian inference for logistic models using Pólya–Gamma latent variables. Journal of the American Statistical Association 108, 1339–1349.
- Effect of model space priors on statistical inference with model uncertainty. The New England Journal of Statistics in Data Science pp. 1–10.
- Laplace power-expected-posterior priors for logistic regression. Bayesian Analysis 1, 1–24.
- Rodríguez, A. (2013). On the jeffreys prior for the multivariate ewens distribution. Statistics & Probability Letters 83, 1539–1546.
- Bayes and empirical-Bayes multiplicity adjustment in the variable-selection problem. The Annals of Statistics pp. 2587–2619.
- Sethuraman, J. (1994). A constructive definition of dirichlet priors. Statistica sinica pp. 639–650.
- Som, A. (2014). Paradoxes and Priors in Bayesian Regression. Ph.D. thesis, The Ohio State University.
- A conditional Lindley paradox in Bayesian linear models. Biometrika 103, 993–999.
- Tipping, M. E. (2001). Sparse Bayesian learning and the relevance vector machine. Journal of Machine Learning Research 1, 211–244.
- Zellner, A. (1986). On assessing prior distributions and Bayesian regression analysis with g-prior distributions. In Bayesian Inference and Decision Techniques: Essays in Honor of Bruno de Finetti, Eds. P. K. Goel & A. Zellner, pp. 233–243. Amsterdam: North-Holland/Elsevier.
- Posterior odds ratios for selected regression hypotheses. Trabajos de Estadística y de Investigaciów Operativa 31, 585–603.