Bayesian Approach to Linear Bayesian Networks (2311.15610v1)
Abstract: This study proposes the first Bayesian approach for learning high-dimensional linear Bayesian networks. The proposed approach iteratively estimates each element of the topological ordering from backward and its parent using the inverse of a partial covariance matrix. The proposed method successfully recovers the underlying structure when Bayesian regularization for the inverse covariance matrix with unequal shrinkage is applied. Specifically, it shows that the number of samples $n = \Omega( d_M2 \log p)$ and $n = \Omega(d_M2 p{2/m})$ are sufficient for the proposed algorithm to learn linear Bayesian networks with sub-Gaussian and 4m-th bounded-moment error distributions, respectively, where $p$ is the number of nodes and $d_M$ is the maximum degree of the moralized graph. The theoretical findings are supported by extensive simulation studies including real data analysis. Furthermore the proposed method is demonstrated to outperform state-of-the-art frequentist approaches, such as the BHLSM, LISTEN, and TD algorithms in synthetic data.
- Bayesian structure learning in graphical models. Journal of Multivariate Analysis, 136:147–162, 2015.
- On causal discovery with an equal-variance assumption. Biometrika, 106(4):973–980, 2019.
- A bayesian approach to learning bayesian networks with local structure. arXiv preprint arXiv:1302.1528, 2013.
- Bayesian regularization for graphical models with unequal shrinkage. Journal of the American Statistical Association, 114(527):1218–1231, 2019.
- A polynomial-time algorithm for learning nonparametric causal graphs. arXiv preprint arXiv:2006.11970, 2020.
- Optimal estimation of gaussian dag models. In International Conference on Artificial Intelligence and Statistics, pages 8738–8757. PMLR, 2022.
- Learning identifiable gaussian bayesian networks in polynomial time and sample complexity. In Advances in Neural Information Processing Systems, pages 6457–6466, 2017.
- Learning linear structural equation models in polynomial time and sample complexity. In Amos Storkey and Fernando Perez-Cruz, editors, Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics, volume 84 of Proceedings of Machine Learning Research, pages 1466–1475, Playa Blanca, Lanzarote, Canary Islands, 09–11 Apr 2018. PMLR.
- Nonlinear causal discovery with additive noise models. In Advances in neural information processing systems, pages 689–696, 2009.
- Estimating high-dimensional directed acyclic graphs with the pc-algorithm. Journal of Machine Learning Research, 8(Mar):613–636, 2007.
- Probabilistic graphical models: principles and techniques. MIT press, 2009.
- Steffen L Lauritzen. Graphical models. Oxford University Press, 1996.
- Bayesian inference for high-dimensional decomposable graphs. Electronic Journal of Statistics, 15(1):1549–1582, 2021.
- High-dimensional learning of linear causal networks via inverse covariance estimation. The Journal of Machine Learning Research, 15(1):3065–3105, 2014.
- Regression by dependence minimization and its application to causal inference in additive noise models. In Proceedings of the 26th annual international conference on machine learning, pages 745–752. ACM, 2009.
- Gunwoong Park. Identifiability of additive noise models using conditional variances. Journal of Machine Learning Research, 21(75):1–34, 2020.
- Gunwoong Park. Computationally efficient learning of gaussian linear structural equation models with equal error variances. Journal of Computational and Graphical Statistics, pages 1–26, 2023.
- Learning high-dimensional gaussian linear structural equation models with heterogeneous error variances. Computational Statistics & Data Analysis, 154:107084, 2021.
- Identifiability of gaussian linear structural equation models with homogeneous and heterogeneous error variances. Journal of the Korean Statistical Society, 49(1):276–292, 2020.
- Learning quadratic variance function (qvf) dag models via overdispersion scoring (ods). Journal of Machine Learning Research, 18(224):1–44, 2018.
- Learning a high-dimensional linear structural equation model via l1-regularized regression. Journal of Machine Learning Research, 22(102):1–41, 2021.
- Identifiability of gaussian structural equation models with equal error variances. Biometrika, 101(1):219–228, 2014.
- Identifiability of causal graphs using functional models. arXiv preprint arXiv:1202.3757, 2012.
- High-dimensional covariance estimation by minimizing ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT-penalized log-determinant divergence. Electronic Journal of Statistics, 5:935–980, 2011.
- The spike-and-slab lasso. Journal of the American Statistical Association, 113(521):431–444, 2018.
- Alberto Roverato. Cholesky decomposition of a hyper inverse wishart matrix. Biometrika, 87(1):99–112, 2000.
- A linear non-Gaussian acyclic model for causal discovery. The Journal of Machine Learning Research, 7:2003–2030, 2006.
- Causation, prediction, and search. MIT press, 2000.
- Geometry of the faithfulness assumption in causal inference. The Annals of Statistics, pages 436–463, 2013.
- Towards scalable bayesian learning of causal dags. Advances in Neural Information Processing Systems, 33:6584–6594, 2020.
- Hao Wang. Scaling it up: Stochastic search structure learning in graphical models. Bayesian Analysis, 10(2):351–377, 2015.
- Causality discovery with additive disturbances: An information-theoretical perspective. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pages 570–585. Springer, 2009a.
- On the identifiability of the post-nonlinear causal model. In Proceedings of the twenty-fifth conference on uncertainty in artificial intelligence, pages 647–655. AUAI Press, 2009b.
- Efficient learning of quadratic variance function directed acyclic graphs via topological layers. arXiv preprint arXiv:2111.01560, 2021.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.