Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
121 tokens/sec
GPT-4o
9 tokens/sec
Gemini 2.5 Pro Pro
47 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Independent Mechanism Analysis and the Manifold Hypothesis (2312.13438v1)

Published 20 Dec 2023 in stat.ML and cs.LG

Abstract: Independent Mechanism Analysis (IMA) seeks to address non-identifiability in nonlinear Independent Component Analysis (ICA) by assuming that the Jacobian of the mixing function has orthogonal columns. As typical in ICA, previous work focused on the case with an equal number of latent components and observed mixtures. Here, we extend IMA to settings with a larger number of mixtures that reside on a manifold embedded in a higher-dimensional than the latent space -- in line with the manifold hypothesis in representation learning. For this setting, we show that IMA still circumvents several non-identifiability issues, suggesting that it can also be a beneficial principle for higher-dimensional observations when the manifold hypothesis holds. Further, we prove that the IMA principle is approximately satisfied with high probability (increasing with the number of observed mixtures) when the directions along which the latent components influence the observations are chosen independently at random. This provides a new and rigorous statistical interpretation of IMA.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (41)
  1. S. Becker and G. E. Hinton. Self-organizing neural network that discovers surfaces in random-dot stereograms. Nature, 355(6356):161–163, 1992.
  2. Representation learning: A review and new perspectives. IEEE transactions on pattern analysis and machine intelligence, 35(8):1798–1828, 2013.
  3. Group invariance principles for causal generative models. In International Conference on Artificial Intelligence and Statistics, pages 557–565. PMLR, 2018.
  4. Bounds on determinants of perturbed diagonal matrices. arXiv preprint arXiv:1401.7084, 2014.
  5. Function classes for identifiable nonlinear independent component analysis. Advances in Neural Information Processing Systems, 35:16946–16961, 2022.
  6. L. Cayton. Algorithms for manifold learning. Univ. of California at San Diego Tech. Rep, 12(1-17):1, 2005.
  7. E. Çinlar. Probability and stochastics, volume 261. Springer, 2011.
  8. P. Comon. Independent component analysis, a new concept? Signal processing, 36(3):287–314, 1994.
  9. Principal manifold flows. arXiv preprint arXiv:2202.07037, 2022.
  10. G. Darmois. Analyse des liaisons de probabilité. In Proc. Int. Stat. Conferences 1947, page 231, 1951.
  11. J. Eriksson and V. Koivunen. Identifiability, separability, and uniqueness of linear ica models. IEEE signal processing letters, 11(7):601–604, 2004.
  12. Testing the manifold hypothesis. Journal of the American Mathematical Society, 29(4):983–1049, 2016.
  13. W. T. Freeman. The generic viewpoint assumption in a framework for visual perception. Nature, 368(6471):542–545, 1994.
  14. The incomplete rosetta stone problem: Identifiability results for multi-view nonlinear ica. In Uncertainty in Artificial Intelligence, pages 217–227. PMLR, 2020.
  15. Independent mechanism analysis, a new concept? In Advances in Neural Information Processing Systems 34 (NeurIPS 2021), Dec. 2021.
  16. H. Hälvä and A. Hyvarinen. Hidden markov nonlinear ica: Unsupervised learning from nonstationary time series. In Conference on Uncertainty in Artificial Intelligence, pages 939–948. PMLR, 2020.
  17. A. Hyvarinen and H. Morioka. Unsupervised feature extraction by time-contrastive learning and nonlinear ica. Advances in neural information processing systems, 29, 2016.
  18. A. Hyvarinen and H. Morioka. Nonlinear ica of temporally dependent stationary sources. In Artificial Intelligence and Statistics, pages 460–469. PMLR, 2017.
  19. A. Hyvärinen and P. Pajunen. Nonlinear independent component analysis: Existence and uniqueness results. Neural networks, 12(3):429–439, 1999.
  20. Independent Component Analysis. John Wiley & Sons, Ltd, 2001.
  21. Nonlinear ica using auxiliary variables and generalized contrastive learning. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 859–868. PMLR, 2019.
  22. Identifiability of latent-variable and structural-equation models: from linear to nonlinear. arXiv preprint arXiv:2302.02672, 2023.
  23. A. Hyvärinen and H. Morioka. Unsupervised Feature Extraction by Time-Contrastive Learning and Nonlinear ICA, 2016.
  24. Y. Ishii. On conharmonic transformations. Tensor, NS, 11:73–80, 1957.
  25. D. Janzing and B. Schölkopf. Causal inference using the algorithmic Markov condition. IEEE Transactions on Information Theory, 56(10):5168–5194, 2010.
  26. Telling cause from effect based on high-dimensional observations. In International Conference on Machine Learning, 2010.
  27. Variational autoencoders and nonlinear ica: A unifying framework. In International Conference on Artificial Intelligence and Statistics, pages 2207–2217. PMLR, 2020.
  28. D. P. Kingma and M. Welling. Auto-encoding variational bayes. In 2nd International Conference on Learning Representations, 2014.
  29. Challenging common assumptions in the unsupervised learning of disentangled representations. In international conference on machine learning, pages 4114–4124. PMLR, 2019.
  30. W. lodzimierz Bryc. Normal distribution characterizations with applications. Lecture Notes in Statistics, 100, 1995.
  31. J. Milnor and D. W. Weaver. Topology from the differentiable viewpoint, volume 21. Princeton university press, 1997.
  32. On the number of linear regions of deep neural networks. In Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada, pages 2924–2932, 2014.
  33. Elements of causal inference: foundations and learning algorithms. The MIT Press, 2017.
  34. Embrace the gap: Vaes perform independent mechanism analysis. Advances in Neural Information Processing Systems, 35:12040–12057, 2022.
  35. Advanced linear algebra, volume 3. Springer, 2005.
  36. W. Rudin et al. Principles of mathematical analysis, volume 3. McGraw-hill New York, 1964.
  37. Probing the robustness of independent mechanism analysis for representation learning. In UAI 2022 Workshop on Causal Representation Learning, 2022.
  38. S. Stepanov and I. Tsyganok. Theorems on conformal mappings of complete riemannian manifolds and their applications. Balkan Journal of Geometry and Its Applications, 22(1):81–86, 2017.
  39. P. Vincent and Y. Bengio. Manifold parzen windows. Advances in neural information processing systems, 15, 2002.
  40. Wikipedia contributors. Rank–nullity theorem — Wikipedia, the free encyclopedia, 2022. [Online; accessed 2-June-2022].
  41. Q. Xi and B. Bloem-Reddy. Indeterminacy in generative models: Characterization and strong identifiability. In International Conference on Artificial Intelligence and Statistics, pages 6912–6939. PMLR, 2023.

Summary

We haven't generated a summary for this paper yet.