AutoGFI: Streamlined Generalized Fiducial Inference for Modern Inference Problems in Models with Additive Errors (2404.08169v2)
Abstract: The concept of fiducial inference was introduced by R. A. Fisher in the 1930s to address the perceived limitations of Bayesian inference, particularly the need for subjective prior distributions in cases with limited prior information. However, Fisher's fiducial approach lost favor due to complications, especially in multi-parameter problems. With renewed interest in fiducial inference in the 2000s, generalized fiducial inference (GFI) emerged as a promising extension of Fisher's ideas, offering new solutions for complex inference challenges. Despite its potential, GFI's adoption has been hindered by demanding mathematical derivations and complex implementation requirements, such as Markov Chain Monte Carlo (MCMC) algorithms. This paper introduces AutoGFI, a streamlined variant of GFI designed to simplify its application across various inference problems with additive noise. AutoGFI's accessibility lies in its simplicity-requiring only a fitting routine-making it a feasible option for a wider range of researchers and practitioners. To demonstrate its efficacy, AutoGFI is applied to three challenging problems: tensor regression, matrix completion, and network cohesion regression. These case studies showcase AutoGFI's competitive performance against specialized solutions, highlighting its potential to broaden the application of GFI in practical domains, ultimately enriching the statistical inference toolkit.
- Beaumont, M. A., Zhang, W., and Balding, D. J. (2002), “Approximate Bayesian computation in population genetics,” Genetics, 162, 2025–2035.
- Bennett, J. and Lanning, S. (2007), “The Netflix prize,” in Proceedings of KDD Cup and Workshop, volume 2007, New York, NY, USA.
- Binkiewicz, N., Vogelstein, J. T., and Rohe, K. (2017), “Covariate-assisted spectral clustering,” Biometrika, 104, 361–377.
- Candès, E. J., Li, X., Ma, Y., and Wright, J. (2011), “Robust principal component analysis?” Journal of the ACM (JACM), 58, 11.
- Candès, E. J. and Plan, Y. (2010), “Matrix completion with noise,” Proceedings of the IEEE, 98, 925–936.
- Candès, E. J. and Recht, B. (2009), “Exact matrix completion via convex optimization,” Foundations of Computational Mathematics, 9, 717.
- Chen, P. and Suter, D. (2004), “Recovering the missing components in a large noisy low-rank matrix: Application to SFM,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 26, 1051–1063.
- Chen, Y., Chi, Y., Fan, J., Ma, C., and Yan, Y. (2020), “Noisy Matrix Completion: Understanding Statistical Guarantees for Convex Relaxation via Nonconvex Optimization,” SIAM Journal on Optimization, 30, 3098–3121.
- Chen, Y., Fan, J., Ma, C., and Yan, Y. (2019), “Inference and uncertainty quantification for noisy matrix completion,” Proceedings of the National Academy of Sciences, 116, 22931–22937.
- Chen, Y. and Wainwright, M. J. (2015), “Fast low-rank estimation by projected gradient descent: General statistical and algorithmic guarantees,” .
- E, L., Hannig, J., Iyer, H., et al. (2008), “Fiducial intervals for variance components in an unbalanced two-component normal mixed linear model,” Journal of the American Statistical Association, 103, 854–865.
- Edlefsen, P. T., Liu, C., and Dempster, A. P. (2009), “Estimating limits from Poisson counting data using Dempster–Shafer analysis,” The Annals of Applied Statistics, 3, 764 – 790, URL https://doi.org/10.1214/00-AOAS223.
- Fisher, R. A. (1922), “On the mathematical foundations of theoretical statistics,” Philosophical Transactions of the Royal Society of London. Series A, Containing Papers of a Mathematical or Physical Character, 222, 309–368.
- — (1925), “Theory of statistical estimation,” in Mathematical Proceedings of the Cambridge Philosophical Society, volume 22, Cambridge University Press, Cambridge University Press.
- — (1930), “Inverse probability,” in Mathematical Proceedings of the Cambridge Philosophical Society, volume 26, Cambridge University Press.
- — (1933), “The concepts of inverse probability and fiducial probability referring to unknown parameters,” Proceedings of the Royal Society of London. Series A, Containing Papers of a Mathematical and Physical Character, 139, 343–348.
- — (1935), “The fiducial argument in statistical inference,” Annals of Eugenics, 6, 391–398.
- Frostig, R., Johnson, M., and Leary, C. (2018), “Compiling machine learning programs via high-level tracing. 2018,” in URL https://mlsys. org/Conferences/doc/2018/146. pdf.
- Gross, D. (2011), “Recovering low-rank matrices from few coefficients in any basis,” IEEE Transactions on Information Theory, 57, 1548–1566.
- Guhaniyogi, R., Qamar, S., and Dunson, D. B. (2017), “Bayesian tensor regression,” The Journal of Machine Learning Research, 18, 2733–2763.
- Guo, W., Kotsia, I., and Patras, I. (2011), “Tensor learning for regression,” IEEE Transactions on Image Processing, 21, 816–827.
- Haldar, J. P. and Liang, Z.-P. (2010), “Spatiotemporal imaging with partially separable functions: A matrix recovery approach,” in Biomedical Imaging: From Nano to Macro, 2010 IEEE International Symposium on, IEEE.
- Han, Y. and Lee, T. C. M. (2022), “Uncertainty Quantification for Sparse Estimation of Spectral Lines,” IEEE Transactions on Signal Processing, 70, 6243–6256.
- Hannig, J. (2009), “ON GENERALIZED FIDUCIAL INFERENCE,” Statistica Sinica, 19, 491–544, URL http://www.jstor.org/stable/24308841.
- — (2013), “Generalized fiducial inference via discretization,” Statistica Sinica, 23, 489–514.
- Hannig, J., Iyer, H., Lai, R. C., and Lee, T. C. M. (2016), “Generalized fiducial inference: A review and new results,” Journal of the American Statistical Association, 111, 1346–1361.
- Hannig, J., Iyer, H., and Patterson, P. (2006), “Fiducial generalized confidence intervals,” Journal of the American Statistical Association, 101, 254–269.
- Hannig, J. and Lee, T. C. M. (2009), “Generalized fiducial inference for wavelet regression,” Biometrika, 96, 847–860.
- Janková, J. and van de Geer, S. (2018), “Inference in high-dimensional graphical models,” in Handbook of Graphical Models, CRC Press, 325–350.
- Keshavan, R. H., Montanari, A., and Oh, S. (2010), “Matrix completion from a few entries,” IEEE Transactions on Information Theory, 56, 2980–2998.
- Kolda, T. G. and Bader, B. W. (2009), “Tensor decompositions and applications,” SIAM Review, 51, 455–500.
- Koltchinskii, V., Lounici, K., and Tsybakov, A. B. (2011), “Nuclear-norm penalization and optimal rates for noisy low-rank matrix completion,” The Annals of Statistics, 39, 2302–2329.
- Lai, R. C., Hannig, J., and Lee, T. C. M. (2015), “Generalized fiducial inference for ultrahigh-dimensional regression,” Journal of the American Statistical Association, 110, 760–772.
- Li, T., Levina, E., Zhu, J., et al. (2019), “Prediction models for network-linked data,” The Annals of Applied Statistics, 13, 132–164.
- Li, X., Xu, D., Zhou, H., and Li, L. (2018), “Tucker tensor regression and neuroimaging analysis,” Statistics in Biosciences, 10, 520–545.
- Liu, Y. and Hannig, J. (2016), “Generalized fiducial inference for binary logistic item response models,” Psychometrika, 81, 290–324.
- Ma, H., Zhou, D., Liu, C., Lyu, M. R., and King, I. (2011), “Recommender systems with social regularization,” in Proceedings of the fourth ACM international conference on Web search and data mining.
- Majumder, A. P. and Hannig, J. (2016), “Higher order asymptotics of Generalized Fiducial Distribution,” .
- Martin, R. and Liu, C. (2013), “Inferential models: A framework for prior-free posterior probabilistic inference,” Journal of the American Statistical Association, 108, 301–313.
- Martin, R., Zhang, J., Liu, C., et al. (2010), “Dempster–Shafer theory and statistical inference with weak beliefs,” Statistical Science, 25, 72–87.
- McNally, R. J., Iyer, H., and Mathew, T. (2003), “Tests for individual and population bioequivalence based on generalized p-values,” Statistics in medicine, 22, 31–53.
- Negahban, S. and Wainwright, M. J. (2012), “Restricted strong convexity and weighted matrix completion: Optimal bounds with noise,” The Journal of Machine Learning Research, 13, 1665–1697.
- Ou-Yang, L., Zhang, X.-F., and Yan, H. (2020), “Sparse regularized low-rank tensor regression with applications in genomic data analysis,” Pattern Recognition, 107, 107516.
- Papadogeorgou, G., Zhang, Z., and Dunson, D. B. (2021), “Soft Tensor Regression.” Journal of Machine Learning Research, 22, 219–1.
- Rao, N., Yu, H.-F., Ravikumar, P. K., and Dhillon, I. S. (2015), “Collaborative Filtering with Graph Information: Consistency and Scalable Methods,” in Cortes, C., Lawrence, N. D., Lee, D. D., Sugiyama, M., and Garnett, R. (editors), Advances in Neural Information Processing Systems 28, Curran Associates, Inc., 2107–2115.
- Rennie, J. D. and Srebro, N. (2005), “Fast maximum margin matrix factorization for collaborative prediction,” in Proceedings of the 22nd International Conference on Machine Learning, ACM.
- Shang, F., Liu, Y., and Cheng, J. (2014), “Generalized higher-order tensor decomposition via parallel ADMM,” in Twenty-Eighth AAAI Conference on Artificial Intelligence.
- Singh, K., Xie, M., Strawderman, W. E., et al. (2007), “Confidence distribution (CD)–distribution estimator of a parameter,” in Complex datasets and inverse problems, Institute of Mathematical Statistics, 132–150.
- Sonderegger, D. L. and Hannig, J. (2014), “Fiducial theory for free-knot splines,” in Contemporary Developments in Statistical Theory, Springer, 155–189.
- Srebro, N. and Jaakkola, T. (2003), “Weighted low-rank approximations,” in Proceedings of the 20th International Conference on Machine Learning (ICML-03).
- Su, Y., Wong, R. K. W., and Lee, T. C. M. (2020), “Network estimation via graphon with node features,” IEEE Transactions on Network Science and Engineering, 7, 2078–2089.
- Tao, C., Nichols, T. E., Hua, X., Ching, C. R., Rolls, E. T., Thompson, P. M., Feng, J., Initiative, A. D. N., et al. (2017), “Generalized reduced rank latent factor regression for high dimensional tensor fields, and neuroimaging-genetic applications,” NeuroImage, 144, 35–57.
- Van de Geer, S., Bühlmann, P., Ritov, Y., and Dezeure, R. (2014), “On asymptotically optimal confidence regions and tests for high-dimensional models,” The Annals of Statistics, 42, 1166–1202, URL http://www.jstor.org/stable/43556319.
- Wang, F., Zhou, L., Tang, L., and Song, P. X. (2021), “Method of contraction-expansion (MOCE) for simultaneous inference in linear models,” The Journal of Machine Learning Research, 22, 8639–8670.
- Xie, M., Singh, K., and Strawderman, W. E. (2011), “Confidence distributions and a unifying framework for meta-analysis,” Journal of the American Statistical Association, 106, 320–333.
- Xie, M.-g. and Singh, K. (2013), “Confidence distribution, the frequentist distribution estimator of a parameter: A review,” International Statistical Review, 81, 3–39.
- Xie, M.-g. and Wang, P. (2022), “Repro Samples Method for Finite- and Large-Sample Inferences,” URL https://arxiv.org/abs/2206.06421.
- Yuchi, H. S., Mak, S., and Xie, Y. (2022), “Bayesian uncertainty quantification for low-rank matrix completion,” Bayesian Analysis, 1, 1–28.
- Zabell, S. L. et al. (1992), “RA Fisher and fiducial argument,” Statistical Science, 7, 369–387.
- Zhang, J. and Liu, C. (2011), “Dempster-Shafer inference with weak beliefs,” Statistica Sinica, 21, 475–494.
- Zhang, Y., Levina, E., Zhu, J., et al. (2016), “Community detection in networks with node features,” Electronic Journal of Statistics, 10, 3153–3178.
- Zhou, H., Li, L., and Zhu, H. (2013), “Tensor Regression with Applications in Neuroimaging Data Analysis,” Journal of the American Statistical Association, 108, 540–552, URL http://www.jstor.org/stable/24246462.