Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
166 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Deep Causal Generative Models with Property Control (2405.16219v1)

Published 25 May 2024 in cs.LG and stat.ML

Abstract: Generating data with properties of interest by external users while following the right causation among its intrinsic factors is important yet has not been well addressed jointly. This is due to the long-lasting challenge of jointly identifying key latent variables, their causal relations, and their correlation with properties of interest, as well as how to leverage their discoveries toward causally controlled data generation. To address these challenges, we propose a novel deep generative framework called the Correlation-aware Causal Variational Auto-encoder (C2VAE). This framework simultaneously recovers the correlation and causal relationships between properties using disentangled latent vectors. Specifically, causality is captured by learning the causal graph on latent variables through a structural causal model, while correlation is learned via a novel correlation pooling algorithm. Extensive experiments demonstrate C2VAE's ability to accurately recover true causality and correlation, as well as its superiority in controllable data generation compared to baseline models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. Causal regularization. arXiv preprint arXiv:1702.02604, 2017.
  2. Invertible residual networks. In International conference on machine learning, pages 573–582. PMLR, 2019.
  3. Regression transformer enables concurrent sequence regression and generation for molecular language modelling. Nature Machine Intelligence, 5(4):432–444, 2023.
  4. Diffusion self-guidance for controllable image generation. Advances in Neural Information Processing Systems, 36, 2024.
  5. Property controllable variational autoencoder via invertible mutual dependence. In International Conference on Learning Representations, 2020.
  6. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017.
  7. Amazon-m2: A multilingual multi-locale shopping session dataset for recommendation and text generation. Advances in Neural Information Processing Systems, 36, 2024.
  8. Variational autoencoders and nonlinear ica: A unifying framework. In International Conference on Artificial Intelligence and Statistics, pages 2207–2217. PMLR, 2020.
  9. Semi-supervised learning with deep generative models. Advances in neural information processing systems, 27, 2014.
  10. Learning latent subspaces in variational autoencoders. Advances in neural information processing systems, 31, 2018.
  11. Disentangling factors of variation using few labels. arXiv preprint arXiv:1905.01258, 2019.
  12. dsprites: Disentanglement testing sprites dataset. https://github.com/deepmind/dsprites-dataset/, 2017.
  13. Causal interpretation of self-attention in pre-trained transformers. Advances in Neural Information Processing Systems, 36, 2024.
  14. Weakly supervised disentangled generative causal representation learning. Journal of Machine Learning Research, 23(241):1–55, 2022.
  15. A linear non-gaussian acyclic model for causal discovery. Journal of Machine Learning Research, 7(10), 2006.
  16. D’ya like dags? a survey on structure learning and causal discovery. ACM Computing Surveys, 55(4):1–36, 2022.
  17. Multi-objective deep data generation with correlated property control. Advances in Neural Information Processing Systems, 35:28889–28901, 2022.
  18. Controllable data generation by deep learning: A review. ACM Computing Surveys, 56(9):1–38, 2024.
  19. Geometric latent diffusion models for 3d molecule generation. In International Conference on Machine Learning, pages 38592–38610. PMLR, 2023.
  20. Causalvae: Disentangled representation learning via neural structural causal models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9593–9602, 2021.
  21. A survey of controllable text generation using transformer-based pre-trained language models. ACM Computing Surveys, 56(3):1–37, 2023.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets