Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Sparsity Principle for Partially Observable Causal Representation Learning (2403.08335v2)

Published 13 Mar 2024 in cs.LG, cs.AI, and stat.ML

Abstract: Causal representation learning aims at identifying high-level causal variables from perceptual data. Most methods assume that all latent causal variables are captured in the high-dimensional observations. We instead consider a partially observed setting, in which each measurement only provides information about a subset of the underlying causal state. Prior work has studied this setting with multiple domains or views, each depending on a fixed subset of latents. Here, we focus on learning from unpaired observations from a dataset with an instance-dependent partial observability pattern. Our main contribution is to establish two identifiability results for this setting: one for linear mixing functions without parametric assumptions on the underlying causal model, and one for piecewise linear mixing functions with Gaussian latent causal variables. Based on these insights, we propose two methods for estimating the underlying causal variables by enforcing sparsity in the inferred representation. Experiments on different simulated datasets and established benchmarks highlight the effectiveness of our approach in recovering the ground-truth latents.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (48)
  1. Identification of partially observed linear causal models: Graphical conditions for the non-gaussian and heterogeneous cases. Advances in Neural Information Processing Systems, 34:22822–22833, 2021.
  2. Weakly supervised representation learning with sparse perturbations. Advances in Neural Information Processing Systems, 35:15516–15528, 2022.
  3. Interventional causal representation learning. In International Conference on Machine Learning, pages 372–407. PMLR, 2023a.
  4. Multi-domain causal representation learning via weak distributional invariances. In Causal Representation Learning Workshop at NeurIPS 2023, 2023b.
  5. Identifying linearly-mixed causal representations from multi-node interventions. arXiv preprint arXiv:2311.02695, 2023.
  6. Weakly supervised causal representation learning. Advances in Neural Information Processing Systems, 35:38319–38331, 2022.
  7. Learning linear causal representations from interventions under general nonlinear mixing. In Advances in Neural Information Processing Systems, 2023.
  8. Triad constraints for learning causal structure of latent variables. Advances in neural information processing systems, 32, 2019.
  9. P. Comon. Independent component analysis, a new concept? Signal Processing, 1994.
  10. Linear Algebra. Pearson Education, 2014. ISBN 9780321998897.
  11. Leveraging sparse and shared feature activations for disentangled representation learning. arXiv preprint arXiv:2304.07939, 2023.
  12. Cooper: a toolkit for lagrangian-based constrained optimization, 2022.
  13. A variational inequality perspective on generative adversarial networks. arXiv preprint arXiv:1802.10551, 2018.
  14. A survey of sparse component analysis for blind source separation: principles, perspectives, and new challenges. In ESANN’06 proceedings-14th European Symposium on Artificial Neural Networks, pages 323–330. d-side publi., 2006.
  15. Global identifiability of $\ell_1$-based dictionary learning via matrix volume optimization. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  16. Unsupervised feature extraction by time-contrastive learning and nonlinear ica. Advances in neural information processing systems, 29, 2016.
  17. Nonlinear ica of temporally dependent stationary sources. In Artificial Intelligence and Statistics, pages 460–469. PMLR, 2017.
  18. Nonlinear ica using auxiliary variables and generalized contrastive learning. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 859–868. PMLR, 2019.
  19. Variational autoencoders and nonlinear ica: A unifying framework. In International Conference on Artificial Intelligence and Statistics, pages 2207–2217. PMLR, 2020.
  20. Learning latent causal graphs via mixture oracles. Advances in Neural Information Processing Systems, 34:18087–18101, 2021.
  21. Identifiability of deep generative models without auxiliary information. Advances in Neural Information Processing Systems, 35:15687–15701, 2022.
  22. Disentanglement via mechanism sparsity regularization: A new principle for nonlinear ica. In Conference on Causal Learning and Reasoning, pages 428–484. PMLR, 2022.
  23. Synergies between disentanglement and sparsity: Generalization and identifiability in multi-task learning. In International Conference on Machine Learning, pages 18171–18206. PMLR, 2023.
  24. Nonparametric partial disentanglement via mechanism sparsity: Sparse actions, interventions and sparse temporal dependencies, 2024.
  25. Citris: Causal identifiability from temporal intervened sequences. In International Conference on Machine Learning, pages 13557–13603. PMLR, 2022.
  26. Biscuit: Causal representation learning from binary interactions. Proceedings of the Thirty-Ninth Conference on Uncertainty in Artificial Intelligence, 2023.
  27. Online dictionary learning for sparse coding. In Proceedings of the 26th annual international conference on machine learning, pages 689–696, 2009.
  28. Identifiable deep generative models via sparse decoding. Transactions on machine learning research, 2022.
  29. Judea Pearl. Causality: Models, Reasoning and Inference. Cambridge University Press, USA, 2nd edition, 2009. ISBN 052189560X.
  30. Causal discovery in heterogeneous environments under the sparse mechanism shift hypothesis. Advances in Neural Information Processing Systems, 35:10904–10917, 2022.
  31. Elements of Causal Inference: Foundations and Learning Algorithms. The MIT Press, 2017. ISBN 0262037319.
  32. Toward causal representation learning. Proceedings of the IEEE, 109(5):612–634, 2021.
  33. Pete Shinners. Pygame, 2011.
  34. Learning the structure of linear latent variable models. Journal of Machine Learning Research, 7(2), 2006.
  35. Causation, Prediction, and Search. MIT Press, Cambridge MA, 2nd edition, 2000.
  36. Linear causal disentanglement via interventions. In 40th International Conference on Machine Learning, 2023.
  37. Unpaired multi-domain causal representation learning. Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  38. Variational sparse coding. In Uncertainty in Artificial Intelligence, pages 690–700. PMLR, 2020.
  39. Score-based causal representation learning from interventions: Nonparametric identifiability. In Causal Representation Learning Workshop at NeurIPS 2023, 2023.
  40. Self-supervised learning with data augmentations provably isolates content from style. Advances in neural information processing systems, 34:16451–16467, 2021.
  41. Nonparametric identifiability of causal representations from unknown interventions. In Advances in Neural Information Processing 36, 2023.
  42. Causal component analysis. In Advances in Neural Information Processing Systems, 2023.
  43. Generalized independent noise condition for estimating latent variable causal graphs. Advances in neural information processing systems, 33:14891–14902, 2020.
  44. Identification of linear non-gaussian latent hierarchical structure. In International Conference on Machine Learning, pages 24370–24387. PMLR, 2022.
  45. Multi-view causal representation learning with partial observability. arXiv preprint arXiv:2311.04056, 2023.
  46. Learning temporally causal latent processes from general temporal data. In International Conference on Learning Representations, 2022.
  47. Identifiability guarantees for causal disentanglement from soft interventions. In Advances in Neural Information Processing Systems, 2023.
  48. Dags with no tears: Continuous optimization for structure learning. Advances in neural information processing systems, 31, 2018.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Danru Xu (3 papers)
  2. Dingling Yao (8 papers)
  3. Sébastien Lachapelle (19 papers)
  4. Perouz Taslakian (31 papers)
  5. Julius von Kügelgen (42 papers)
  6. Francesco Locatello (92 papers)
  7. Sara Magliacane (26 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com