Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 83 tok/s
Gemini 2.5 Pro 34 tok/s Pro
GPT-5 Medium 24 tok/s Pro
GPT-5 High 21 tok/s Pro
GPT-4o 130 tok/s Pro
Kimi K2 207 tok/s Pro
GPT OSS 120B 460 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

MissNODAG: Differentiable Cyclic Causal Graph Learning from Incomplete Data (2410.18918v1)

Published 24 Oct 2024 in stat.ML and cs.LG

Abstract: Causal discovery in real-world systems, such as biological networks, is often complicated by feedback loops and incomplete data. Standard algorithms, which assume acyclic structures or fully observed data, struggle with these challenges. To address this gap, we propose MissNODAG, a differentiable framework for learning both the underlying cyclic causal graph and the missingness mechanism from partially observed data, including data missing not at random. Our framework integrates an additive noise model with an expectation-maximization procedure, alternating between imputing missing values and optimizing the observed data likelihood, to uncover both the cyclic structures and the missingness mechanism. We demonstrate the effectiveness of MissNODAG through synthetic experiments and an application to real-world gene perturbation data.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (71)
  1. Structure learning for cyclic linear causal models. In Conference on Uncertainty in Artificial Intelligence, pages 999–1008. PMLR.
  2. Invertible residual networks. In International Conference on Machine Learning, pages 573–582. PMLR.
  3. Identification in missing data models represented by directed acyclic graphs. In Uncertainty in artificial intelligence, pages 1149–1158. PMLR.
  4. Differentiable causal discovery under unmeasured confounding. In International Conference on Artificial Intelligence and Statistics, pages 2314–2322. PMLR.
  5. Bollen, K. A. (1989). Structural equations with latent variables, volume 210. John Wiley & Sons.
  6. Carter, R. L. (2006). Solutions for missing data in structural equation modeling. Research & Practice in Assessment, 1:4–7.
  7. A penalized em algorithm incorporating missing data mechanism for gaussian parameter estimation. Biometrics, 70(2):312–322.
  8. Residual flows for invertible generative modeling. Advances in Neural Information Processing Systems, 32.
  9. Computation of maximum likelihood estimates in cyclic structural equation models. The Annals of Statistics, 47(2):663 – 690.
  10. Multimodal pooled Perturb-CITE-seq screens in patient models define mechanisms of cancer immune evasion. Nature genetics, 53(3):332–341.
  11. Systematic discovery and perturbation of regulatory genes in human T cells reveals the architecture of immune networks. Nature Genetics, pages 1–12.
  12. Friedman, N. (1998). The bayesian structural em algorithm. In Conference on Uncertainty in Artificial Intelligence.
  13. Structure learning under missing data. In International conference on probabilistic graphical models, pages 121–132. PMLR.
  14. Missdag: Causal discovery in the presence of missing data with continuous additive noise models. Advances in Neural Information Processing Systems, 35:5024–5038.
  15. Mining for equitable health: Assessing the impact of missing data in electronic health records. Journal of biomedical informatics, 139:104269.
  16. Characterizing distribution equivalence and structure learning for cyclic and acyclic directed graphs. In International Conference on Machine Learning, pages 3494–3504. PMLR.
  17. Sufficient identification conditions and semiparametric estimation under missing not at random mechanisms. In Uncertainty in Artificial Intelligence, pages 777–787. PMLR.
  18. Hall, B. C. (2013). Lie Groups, Lie Algebras, and Representations, pages 333–366. Springer New York, New York, NY.
  19. Characterization and greedy learning of interventional markov equivalence classes of directed acyclic graphs. The Journal of Machine Learning Research, 13(1):2409–2464.
  20. Invariant causal prediction for nonlinear models. Journal of Causal Inference, 6(2).
  21. Estimation rates for sparse linear cyclic causal models. In Peters, J. and Sontag, D., editors, Proceedings of the 36th Conference on Uncertainty in Artificial Intelligence (UAI), volume 124 of Proceedings of Machine Learning Research, pages 1169–1178. PMLR.
  22. Hutchinson, M. F. (1989). A stochastic estimator of the trace of the influence matrix for Laplacian smoothing splines. Communications in Statistics-Simulation and Computation, 18(3):1059–1076.
  23. Learning linear cyclic causal models with latent variables. The Journal of Machine Learning Research, 13(1):3387–3439.
  24. Causal inference in statistics, social, and biomedical sciences. Cambridge University Press.
  25. Categorical reparameterization with Gumbel-Softmax. arXiv preprint arXiv:1611.01144.
  26. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
  27. Probabilistic graphical models: principles and techniques. MIT press.
  28. Miracle: Causally-aware imputation via learning missing data mechanisms. Advances in Neural Information Processing Systems, 34:23806–23817.
  29. Discovering cyclic causal models by independent components analysis. In Proceedings of the Twenty-Fourth Conference on Uncertainty in Artificial Intelligence, UAI’08, page 366–374, Arlington, Virginia, USA. AUAI Press.
  30. Scaling structural learning with NO-BEARS to infer causal transcriptome networks. In Pacific Symposium on Biocomputing 2020, pages 391–402. World Scientific.
  31. Learning from incomplete data with generative adversarial networks. In International Conference on Learning Representations.
  32. Statistical analysis with missing data, volume 793. John Wiley & Sons.
  33. Large-scale differentiable causal discovery of factor graphs. Advances in Neural Information Processing Systems, 35:19290–19303.
  34. Multivariate time series imputation with generative adversarial networks. Advances in neural information processing systems, 31.
  35. Meek, C. (1997). Graphical Models: Selecting causal and statistical models. PhD thesis, Carnegie Mellon University.
  36. Graphical models for processing missing data. Journal of the American Statistical Association, 116(534):1023–1037.
  37. Graphical models for inference with missing data. In Burges, C., Bottou, L., Welling, M., Ghahramani, Z., and Weinberger, K., editors, Advances in Neural Information Processing Systems, volume 26. Curran Associates, Inc.
  38. Cyclic causal discovery from continuous equilibrium data. In Uncertainty in Artificial Intelligence.
  39. Missing data imputation using optimal transport. In International Conference on Machine Learning, pages 7130–7140. PMLR.
  40. On testability and goodness of fit tests in missing data models. In Uncertainty in Artificial Intelligence, pages 1467–1477. PMLR.
  41. Full law identification in graphical models of missing data: Completeness results. In International conference on machine learning, pages 7153–7163. PMLR.
  42. Causal and counterfactual views of missing data models. arXiv preprint arXiv:2210.05558.
  43. On the role of sparsity and DAG constraints for learning linear dags. Advances in Neural Information Processing Systems, 33:17943–17954.
  44. Masked gradient-based causal structure learning. In Proceedings of the 2022 SIAM International Conference on Data Mining (SDM), pages 424–432. SIAM.
  45. Pearl, J. (2009a). Causality. Cambridge University Press, 2 edition.
  46. Pearl, J. (2009b). Causality: Models, Reasoning, and Inference. Cambridge University Press, 2 edition.
  47. Richardson, T. (1996). A discovery algorithm for directed cyclic graphs. In Proceedings of the Twelfth international conference on Uncertainty in artificial intelligence, pages 454–461.
  48. Rudin, W. (1953). Principles of Mathematical Analysis. McGraw-Hill Book Company, Inc., New York-Toronto-London.
  49. Causal protein-signaling networks derived from multiparameter single-cell data. Science, 308(5721):523–529.
  50. Anchored causal inference in the presence of measurement error. In Conference on uncertainty in artificial intelligence, pages 619–628. PMLR.
  51. Review of inverse probability weighting for dealing with missing data. Statistical methods in medical research, 22(3):278–295.
  52. Learning module networks. Journal of Machine Learning Research, 6(4).
  53. Nodags-flow: Nonlinear cyclic causal structure learning. In Proceedings of The 26th International Conference on Artificial Intelligence and Statistics, volume 206 of Proceedings of Machine Learning Research, pages 6371–6387. PMLR.
  54. Singh, M. (1997). Learning bayesian networks from incomplete data. AAAI/IAAI, 1001:534–539.
  55. Consistency guarantees for permutation-based causal inference algorithms. arXiv preprint arXiv:1702.03530.
  56. Causation, prediction, and search. MIT press.
  57. Missforest—non-parametric missing value imputation for mixed-type data. Bioinformatics, 28(1):112–118.
  58. Fast causal inference with non-random missingness by test-wise deletion. International journal of data science and analytics, 6:47–62.
  59. Encoding dependence in bayesian causal networks. Frontiers in Environmental Science, 4:84.
  60. Constraint-based causal discovery from multiple interventions over overlapping variable sets. The Journal of Machine Learning Research, 16(1):2147–2205.
  61. The max-min hill-climbing bayesian network structure learning algorithm. Machine learning, 65(1):31–78.
  62. Causal discovery in the presence of missing data. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 1762–1770. PMLR.
  63. Efficient algorithms for bayesian network parameter learning from incomplete data. In Proceedings of the Thirty-First Conference on Uncertainty in Artificial Intelligence, UAI’15, page 161–170, Arlington, Virginia, USA. AUAI Press.
  64. Causal discovery from incomplete data: a deep learning approach. arXiv preprint arXiv:2001.05343.
  65. Permutation-based causal inference algorithms with interventions. Advances in Neural Information Processing Systems, 30.
  66. Multiple imputation using chained equations: issues and guidance for practice. Statistics in medicine, 30(4):377–399.
  67. Wu, C. F. J. (1983). On the Convergence Properties of the EM Algorithm. The Annals of Statistics, 11(1):95 – 103.
  68. DAG-GNN: DAG structure learning with graph neural networks. In International Conference on Machine Learning, pages 7154–7163. PMLR.
  69. Integrated systems approach identifies genetic nodes and networks in late-onset Alzheimer’s disease. Cell, 153(3):707–720.
  70. DAGs with NO TEARS: Continuous optimization for structure learning. In Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 31.
  71. Learning sparse nonparametric DAGs. In Chiappa, S. and Calandra, R., editors, Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics, volume 108, pages 3414–3425.

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 post and received 0 likes.