Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Causal Discovery from Poisson Branching Structural Causal Model Using High-Order Cumulant with Path Analysis (2403.16523v1)

Published 25 Mar 2024 in stat.ML, cs.AI, and cs.LG

Abstract: Count data naturally arise in many fields, such as finance, neuroscience, and epidemiology, and discovering causal structure among count data is a crucial task in various scientific and industrial scenarios. One of the most common characteristics of count data is the inherent branching structure described by a binomial thinning operator and an independent Poisson distribution that captures both branching and noise. For instance, in a population count scenario, mortality and immigration contribute to the count, where survival follows a Bernoulli distribution, and immigration follows a Poisson distribution. However, causal discovery from such data is challenging due to the non-identifiability issue: a single causal pair is Markov equivalent, i.e., $X\rightarrow Y$ and $Y\rightarrow X$ are distributed equivalent. Fortunately, in this work, we found that the causal order from $X$ to its child $Y$ is identifiable if $X$ is a root vertex and has at least two directed paths to $Y$, or the ancestor of $X$ with the most directed path to $X$ has a directed path to $Y$ without passing $X$. Specifically, we propose a Poisson Branching Structure Causal Model (PB-SCM) and perform a path analysis on PB-SCM using high-order cumulants. Theoretical results establish the connection between the path and cumulant and demonstrate that the path information can be obtained from the cumulant. With the path information, causal order is identifiable under some graphical conditions. A practical algorithm for learning causal structure under PB-SCM is proposed and the experiments demonstrate and verify the effectiveness of the proposed method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. First-order integer-valued autoregressive (INAR (1)) process. Journal of Time Series Analysis, 8(3): 261–275.
  2. Self: structural equational likelihood framework for causal discovery. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32.
  3. THPs: Topological Hawkes Processes for Learning Causal Structure on Event Sequences. IEEE Transactions on Neural Networks and Learning Systems.
  4. Chickering, D. M. 2002. Optimal Structure Identification with Greedy Search. Journal of machine learning research, 3(Nov): 507–554.
  5. Bayesian causal structural learning with zero-inflated poisson bayesian networks. Advances in neural information processing systems, 33: 5887–5897.
  6. Introduction to algorithms. MIT press.
  7. An introduction to the bootstrap. CRC press.
  8. Review of causal discovery methods based on graphical models. Frontiers in genetics, 10: 524.
  9. McKenzie, E. 1985. Some simple models for discrete variate time series 1. JAWRA Journal of the American Water Resources Association, 21(4): 645–650.
  10. Ordinal causal discovery. In Uncertainty in Artificial Intelligence, 1530–1540. PMLR.
  11. Learning large-scale poisson dag models based on overdispersion scoring. Advances in neural information processing systems, 28.
  12. Learning Quadratic Variance Function (QVF) DAG Models via OverDispersion Scoring (ODS). Journal of Machine Learning Research, 18(224): 1–44.
  13. Pearl, J. 2009. Causality. Cambridge university press.
  14. Structural Hawkes Processes for Learning Causal Structure from Discrete-Time Event Sequences. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI-23, 5702–5710.
  15. Causation, prediction, and search. MIT press.
  16. Causal inference in the presence of latent variables and selection bias. In Proceedings of the Eleventh conference on Uncertainty in artificial intelligence, 499–506.
  17. Discrete analogues of self-decomposability and stability. The Annals of Probability, 893–899.
  18. The max-min hill-climbing Bayesian network structure learning algorithm. Machine learning, 65: 31–78.
  19. Weiß, C. H. 2018. An introduction to discrete-valued time series. John Wiley & Sons.
  20. Diagnosing and modeling extra-binomial variation for time-dependent counts. Applied Stochastic Models in Business and Industry, 30(5): 588–608.
  21. Binomial subsampling. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 462(2068): 1181–1195.
  22. Learning causality and causality-related learning: some recent progress. National science review, 5(1): 26–29.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com