Federated Causality Learning with Explainable Adaptive Optimization (2312.05540v1)
Abstract: Discovering the causality from observational data is a crucial task in various scientific domains. With increasing awareness of privacy, data are not allowed to be exposed, and it is very hard to learn causal graphs from dispersed data, since these data may have different distributions. In this paper, we propose a federated causal discovery strategy (FedCausal) to learn the unified global causal graph from decentralized heterogeneous data. We design a global optimization formula to naturally aggregate the causal graphs from client data and constrain the acyclicity of the global graph without exposing local data. Unlike other federated causal learning algorithms, FedCausal unifies the local and global optimizations into a complete directed acyclic graph (DAG) learning process with a flexible optimization objective. We prove that this optimization objective has a high interpretability and can adaptively handle homogeneous and heterogeneous data. Experimental results on synthetic and real datasets show that FedCausal can effectively deal with non-independently and identically distributed (non-iid) data and has a superior performance.
- Ordering-based causal structure learning in the presence of latent variables. In AISTATS, 4098–4108.
- CAM: Causal additive models, high-dimensional order search and penalized regression. The Annals of Statistics, 42(6): 2526–2556.
- A limited memory algorithm for bound constrained optimization. SIAM Journal on Scientific Computing, 16(5): 1190–1208.
- Secure federated matrix factorization. IEEE Intelligent Systems, 36(5): 11–20.
- Chickering, D. M. 2002. Optimal structure identification with greedy search. JMLR, 3(11): 507–554.
- FedDAG: Federated DAG Structure Learning. TMLR.
- Learning Bayesian network structure from distributed homogeneous data. In In ACIS International Conference on Software Engineering, Artificial Intelligence, Networking, and Parallel/Distributed Computing, volume 3, 250–254. IEEE.
- Daring: Differentiable causal discovery with residual independence. In KDD, 596–605.
- Probabilistic graphical models: principles and techniques. MIT press.
- Gradient-Based Neural DAG Learning. In ICLR.
- Directed Acyclic Graph Learning on Attributed Heterogeneous Network. IEEE Transactions on Knowledge and Data Engineering.
- Communication-efficient learning of deep networks from decentralized data. In AISTATS, 1273–1282.
- Distributed Bayesian network structure learning. In ISIE, 1607–1611.
- Nemirovsky, A. 1999. Optimization II. Numerical methods for nonlinear continuous optimization.
- Towards federated bayesian network structure learning with continuous optimization. In AISTATS, 8095–8111.
- Masked gradient-based causal structure learning. In SDM, 424–432.
- Pearl, J. 2009. Causality. Cambridge university press.
- Elements of causal inference: foundations and learning algorithms. MIT Press.
- Causal protein-signaling networks derived from multiparameter single-cell data. Science, 308(5721): 523–529.
- Shimizu, S. 2012. Joint estimation of linear non-Gaussian acyclic models. Neurocomputing, 81: 104–107.
- Causation, prediction, and search. MIT Press.
- Causal inference in the presence of latent variables and selection bias. arXiv preprint arXiv:1302.4983.
- Integrating locally learned causal structures with overlapping variables. In NeurIPS, 1665–1672.
- Constraint-based causal discovery from multiple interventions over overlapping variable sets. JMLR, 16(1): 2147–2205.
- The max-min hill-climbing Bayesian network structure learning algorithm. Mach. Learn., 65(1): 31–78.
- Federated causal inference in heterogeneous observational data. arXiv preprint arXiv:2107.11732.
- Distributed Learning of Generalized Linear Causal Networks. arXiv preprint arXiv:2201.09194.
- DAG-GNN: DAG structure learning with graph neural networks. In ICML, 7154–7163.
- Yuan, M. 2011. On the identifiability of additive index models. Statistica Sinica, 1901–1911.
- DAGs with NO TEARS: continuous optimization for structure learning. In NeurIPS, 9492–9503.
- Learning sparse nonparametric dags. In AISTATS, 3414–3425.