Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Modular Learning of Deep Causal Generative Models for High-dimensional Causal Inference (2401.01426v2)

Published 2 Jan 2024 in cs.LG, cs.AI, cs.IT, math.IT, stat.ME, and stat.ML

Abstract: Sound and complete algorithms have been proposed to compute identifiable causal queries using the causal structure and data. However, most of these algorithms assume accurate estimation of the data distribution, which is impractical for high-dimensional variables such as images. On the other hand, modern deep generative architectures can be trained to sample from high-dimensional distributions. However, training these networks are typically very costly. Thus, it is desirable to leverage pre-trained models to answer causal queries using such high-dimensional data. To address this, we propose modular training of deep causal generative models that not only makes learning more efficient, but also allows us to utilize large, pre-trained conditional generative models. To the best of our knowledge, our algorithm, Modular-DCM is the first algorithm that, given the causal structure, uses adversarial training to learn the network weights, and can make use of pre-trained models to provably sample from any identifiable causal query in the presence of latent confounders. With extensive experiments on the Colored-MNIST dataset, we demonstrate that our algorithm outperforms the baselines. We also show our algorithm's convergence on the COVIDx dataset and its utility with a causal invariant prediction problem on CelebA-HQ.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (38)
  1. Partial identification of treatment effects with implicit generative models. Advances in Neural Information Processing Systems, 35:22816–22829.
  2. Bard (2023). Bard - chat based ai tool from google, powered by palm 2. https://bard.google.com/chat. Accessed: Sep 26, 2023.
  3. Causal inference by surrogate experiments: z-identifiability. arXiv preprint arXiv:1210.4842.
  4. Causal inference by surrogate experiments: Z-identifiability. In Proceedings of the Twenty-Eighth Conference on Uncertainty in Artificial Intelligence, UAI’12, page 113–120, Arlington, Virginia, USA. AUAI Press.
  5. Estimating the effects of continuous-valued interventions using generative adversarial networks. Advances in Neural Information Processing Systems, 33:16434–16445.
  6. Morpho-mnist: quantitative assessment and diagnostics for representation learning. Journal of Machine Learning Research, 20(178):1–29.
  7. ChatGPT (2022). Chatgpt: A large language model trained by openai. https://openai.com/blog/chatgpt/. Accessed: Sep 26, 2023.
  8. Nested counterfactual identification from arbitrary surrogate experiments. Advances in Neural Information Processing Systems, 34:6856–6867.
  9. DALL-E (2022). Dall-e 2 is an ai system that can create realistic images and art from a description in natural language. https://openai.com/dall-e-2. Accessed: Sep 26, 2023.
  10. Cxr-acgan: Auxiliary classifier gan (ac-gan) for conditional generation of chest x-ray images (pneumonia, covid-19 and healthy patients) for the purpose of data augmentation. https://github.com/giocoal/CXR-ACGAN-chest-xray-generator-covid19-pneumonia.
  11. Improved training of wasserstein gans. Advances in neural information processing systems, 30.
  12. Classifier-free diffusion guidance. arXiv preprint arXiv:2207.12598.
  13. Pearl’s calculus of intervention is complete. arXiv preprint arXiv:1206.6831.
  14. Causal identification under markov equivalence: Completeness results. In International Conference on Machine Learning, pages 2981–2989. PMLR.
  15. Categorical reparameterization with gumbel-softmax. In International Conference on Learning Representations.
  16. Causalgan: Learning causal implicit generative models with adversarial training. In International Conference on Learning Representations.
  17. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324.
  18. General identifiability with arbitrary surrogate experiments. In Uncertainty in artificial intelligence, pages 389–398. PMLR.
  19. Causal effect inference with deep latent-variable models. Advances in neural information processing systems, 30.
  20. Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784.
  21. Countergan: generating realistic counterfactuals with residual generative adversarial nets. arXiv preprint arXiv:2009.05199.
  22. Deep structural causal models for tractable counterfactual inference. Advances in Neural Information Processing Systems, 33:857–869.
  23. Pearl, J. (1993). [bayesian analysis in expert systems]: comment: graphical models, causality and intervention. Statistical Science, 8(3):266–269.
  24. Pearl, J. (1995). Causal diagrams for empirical research. Biometrika, 82(4):669–688.
  25. Pearl, J. (2009). Causality. Cambridge university press.
  26. Causal protein-signaling networks derived from multiparameter single-cell data. Science, 308(5721):523–529.
  27. Bayesian networks: with examples in R. Chapman and Hall/CRC.
  28. What counterfactuals can be tested. In Proceedings of the Twenty-Third Conference on Uncertainty in Artificial Intelligence, pages 352–359.
  29. Complete identification methods for the causal hierarchy. Journal of Machine Learning Research, 9:1941–1979.
  30. Effects of treatment on the treated: Identification and generalization. In Proceedings of the 25th Conference on Uncertainty in Artificial Intelligence, UAI 2009, pages 514–521. AUAI Press.
  31. A characterization of interventional distributions in semi-markovian causal models. In AAAI, pages 1239–1244.
  32. A general identification condition for causal effects. eScholarship, University of California.
  33. Decaf: Generating fair synthetic data using causally-aware generative networks. Advances in Neural Information Processing Systems, 34:22221–22233.
  34. Covid-net: a tailored deep convolutional neural network design for detection of covid-19 cases from chest x-ray images. Scientific Reports, 10(1):19549.
  35. The causal-neural connection: Expressiveness, learnability, and inference. Advances in Neural Information Processing Systems, 34:10823–10836.
  36. Neural causal models for counterfactual identification and estimation. In The Eleventh International Conference on Learning Representations.
  37. Achieving causal fairness through generative adversarial networks. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence.
  38. Treatment effect estimation with disentangled latent factors. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 10923–10930.
Citations (1)

Summary

We haven't generated a summary for this paper yet.