Papers
Topics
Authors
Recent
Search
2000 character limit reached

Covariance Regression with High-Dimensional Predictors

Published 10 Apr 2024 in stat.ME, math.ST, and stat.TH | (2404.06701v1)

Abstract: In the high-dimensional landscape, addressing the challenges of covariance regression with high-dimensional covariates has posed difficulties for conventional methodologies. This paper addresses these hurdles by presenting a novel approach for high-dimensional inference with covariance matrix outcomes. The proposed methodology is illustrated through its application in elucidating brain coactivation patterns observed in functional magnetic resonance imaging (fMRI) experiments and unraveling complex associations within anatomical connections between brain regions identified through diffusion tensor imaging (DTI). In the pursuit of dependable statistical inference, we introduce an integrative approach based on penalized estimation. This approach combines data splitting, variable selection, aggregation of low-dimensional estimators, and robust variance estimation. It enables the construction of reliable confidence intervals for covariate coefficients, supported by theoretical confidence levels under specified conditions, where asymptotic distributions are provided. Through various types of simulation studies, the proposed approach performs well for covariance regression in the presence of high-dimensional covariates. This innovative approach is applied to the Lifespan Human Connectome Project (HCP) Aging Study, which aims to uncover a typical aging trajectory and variations in the brain connectome among mature and older adults. The proposed approach effectively identifies brain networks and associated predictors of white matter integrity, aligning with established knowledge of the human brain.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (44)
  1. Small-world brain networks revisited. The Neuroscientist, 23(5):499–516.
  2. Inference on treatment effects after selection among high-dimensional controls. The Review of Economic Studies, 81(2):608–650.
  3. Statistics for high-dimensional data: methods, theory and applications. Springer Science & Business Media.
  4. Analyzing bagging. The annals of Statistics, 30(4):927–961.
  5. The economy of brain network organization. Nature Reviews Neuroscience, 13(5):336.
  6. An automated labeling system for subdividing the human cerebral cortex on MRI scans into gyral based regions of interest. NeuroImage, 31(3):968–980.
  7. Efron, B. (2014). Estimation and accuracy after model selection. Journal of the American Statistical Association, 109(507):991–1007.
  8. Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American statistical Association, 96(456):1348–1360.
  9. Sure independence screening for ultrahigh dimensional feature space. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 70(5):849–911.
  10. Estimation and inference for high dimensional generalized linear models: A splitting and smoothing approach. The Journal of Machine Learning Research, 22(1):2681–2712.
  11. The minimal preprocessing pipelines for the Human Connectome Project. NeuroImage, 80:105–124.
  12. Fragility and volatility of structural hubs in the human connectome. Nature Neuroscience, 21(8):1107–1116.
  13. Interpretable whole-brain prediction analysis with GraphNet. NeuroImage, 72:304–321.
  14. Tau deposition and structural connectivity demonstrate differential association patterns with neurocognitive tests. Brain Imaging and Behavior, 16(2):702–714.
  15. Hebb, D. O. (2005). The organization of behavior: A neuropsychological theory. Psychology Press.
  16. Confidence intervals and hypothesis testing for high-dimensional regression. The Journal of Machine Learning Research, 15(1):2869–2909.
  17. Tractography: where do we go from here? Brain Connectivity, 1(3):169–183.
  18. High-dimensional covariance regression with application to co-expression qtl detection. arXiv preprint arXiv:2404.02093.
  19. Investigating the functional heterogeneity of the default mode network using coordinate-based meta-analytic modeling. Journal of Neuroscience, 29(46):14496–14505.
  20. Exact post-selection inference, with application to the lasso. The Annals of Statistics, 44(3):907–927.
  21. On model selection consistency of regularized M-estimators. Electronic Journal of Statistics, 9(1):608–642.
  22. Fractionating the default mode network: distinct contributions of the ventral and dorsal posterior cingulate cortex to cognitive control. Journal of Neuroscience, 31(9):3217–3224.
  23. Individual variability in functional connectivity architecture of the human brain. Neuron, 77(3):586–595.
  24. A unified framework for high-dimensional analysis of m𝑚mitalic_m-estimators with decomposable regularizers. Statistical Science, 27(4):538–557.
  25. eneral theory of hypothesis tests and confidence regions for sparse high dimensional models. The Annals of Statistics, 45(1):158–195.
  26. Multivariate heteroscedasticity models for functional brain connectivity. Frontiers in Neuroscience, 11:696.
  27. Advances in functional and structural MR image analysis and implementation as FSL. NeuroImage, 23:S208–S219.
  28. Convex and non-convex approaches for statistical inference with class-conditional noisy labels. The Journal of Machine Learning Research, 21(1):6805–6862.
  29. Sporns, O. (2007). Brain connectivity. Scholarpedia, 2(10):4695.
  30. Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society: Series B (Methodological), 58(1):267–288.
  31. Sparsity and smoothness via the fused lasso. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 67(1):91–108.
  32. The solution path of the generalized lasso. The Annals of Statistics, 39(3):1335–1371.
  33. On asymptotically optimal confidence regions and tests for high-dimensional models. The Annals of Statistics, 42(3):1166–1202.
  34. Vershynin, R. (2010). Introduction to the non-asymptotic analysis of random matrices. arXiv preprint arXiv:1011.3027.
  35. Estimation and inference of heterogeneous treatment effects using random forests. Journal of the American Statistical Association, 113(523):1228–1242.
  36. High dimensional variable selection. Annals of statistics, 37(5A):2178.
  37. A revisit to de-biased lasso for generalized linear models. arXiv preprint arXiv:2006.12778.
  38. Zhang, C.-H. (2010). Nearly unbiased variable selection under minimax concave penalty. The Annals of Statistics, pages 894–942.
  39. Confidence intervals for low dimensional parameters in high dimensional linear models. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 76(1):217–242.
  40. Covariate assisted principal regression for covariance matrix outcomes. Biostatistics.
  41. Covariate assisted principal regression for covariance matrix outcomes. Biostatistics, 22(3):629–645.
  42. Zou, H. (2006). The adaptive lasso and its oracle properties. Journal of the American Statistical Association, 101(476):1418–1429.
  43. Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 67(2):301–320.
  44. Covariance regression analysis. Journal of the American Statistical Association, 112(517):266–281.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (3)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.