Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
121 tokens/sec
GPT-4o
9 tokens/sec
Gemini 2.5 Pro Pro
47 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Identifying Patient-Specific Root Causes with the Heteroscedastic Noise Model (2205.13085v2)

Published 25 May 2022 in stat.ML, cs.LG, stat.AP, and stat.ME

Abstract: Complex diseases are caused by a multitude of factors that may differ between patients even within the same diagnostic category. A few underlying root causes may nevertheless initiate the development of disease within each patient. We therefore focus on identifying patient-specific root causes of disease, which we equate to the sample-specific predictivity of the exogenous error terms in a structural equation model. We generalize from the linear setting to the heteroscedastic noise model where $Y = m(X) + \varepsilon\sigma(X)$ with non-linear functions $m(X)$ and $\sigma(X)$ representing the conditional mean and mean absolute deviation, respectively. This model preserves identifiability but introduces non-trivial challenges that require a customized algorithm called Generalized Root Causal Inference (GRCI) to extract the error terms correctly. GRCI recovers patient-specific root causes more accurately than existing alternatives.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (42)
  1. A linear non-Gaussian acyclic model for causal discovery. Journal of Machine Learning Research. 2006;7(10).
  2. Strobl EV, Lasko TA. Identifying Patient-Specific Root Causes of Disease. arXiv preprint arXiv:220511627. 2022.
  3. Nonlinear causal discovery with additive noise models. Advances in Neural Information Processing Systems. 2008;21.
  4. Zhang K, Hyvärinen A. On the Identifiability of the Post-Nonlinear Causal Model. In: 25th Conference on Uncertainty in Artificial Intelligence (UAI 2009). AUAI Press; 2009. p. 647-55.
  5. Independence properties of directed Markov fields. Networks. 1990;20(5):491-505.
  6. Causal Discovery with Continuous Additive Noise Models. Journal of Machine Learning Research. 2014;15:2009-53.
  7. Causal Inference with Heteroscedastic Noise Models. Proceedings of the AAAI Workshop on Information Theoretic Causal Inference and Discovery. 2022.
  8. FOM: Fourth-order moment based causal direction identification on the heteroscedastic data. Neural Networks. 2020;124:193-201.
  9. Distinguishing cause from effect using quantiles: Bivariate quantile causal discovery. In: International Conference on Machine Learning. PMLR; 2020. p. 9311-23.
  10. Causal inference via kernel deviance measures. Advances in Neural Information Processing Systems. 2018;31.
  11. Liu F, Chan LW. Causal inference on multidimensional data using free probability theory. IEEE transactions on neural networks and learning systems. 2017;29(7):3188-98.
  12. Causal structure based root cause analysis of outliers. arXiv preprint arXiv:191202724. 2019.
  13. Why did the distribution change? In: International Conference on Artificial Intelligence and Statistics. PMLR; 2021. p. 1666-74.
  14. Causal structure-based root cause analysis of outliers. In: International Conference on Machine Learning. PMLR; 2022. p. 2357-69.
  15. Uemura K, Shimizu S. Estimation of post-nonlinear causal models using autoencoding structure. In: ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE; 2020. p. 3312-6.
  16. Breiman L, Friedman JH. Estimating optimal transformations for multiple regression and correlation. Journal of the American statistical Association. 1985;80(391):580-98.
  17. On estimation of functional causal models: general results and application to the post-nonlinear causal model. ACM Transactions on Intelligent Systems and Technology (TIST). 2015;7(2):1-22.
  18. Rank-Based Causal Discovery for Post-Nonlinear Models. 26th International Conference on Artificial Intelligence and Statistics (AISTATS). 2023.
  19. Janzing D, Steudel B. Justifying additive noise model-based causal discovery via algorithmic information theory. Open Systems & Information Dynamics. 2010;17(02):189-212.
  20. Janzing D, Schölkopf B. Causal inference using the algorithmic Markov condition. IEEE Transactions on Information Theory. 2010;56(10):5168-94.
  21. Probabilistic latent variable models for distinguishing between cause and effect. Advances in Neural Information Processing Systems. 2010;23.
  22. Information-geometric approach to inferring causal directions. Artificial Intelligence. 2012;182:1-31.
  23. Justifying information-geometric causal inference. In: Measures of Complexity. Springer; 2015. p. 253-65.
  24. Identifiability of causal graphs using functional models. In: 27th Conference on Uncertainty in Artificial Intelligence (UAI 2011). AUAI Press; 2011. p. 589-98.
  25. Consistent individualized feature attribution for tree ensembles. arXiv preprint arXiv:180203888. 2018.
  26. Backtracking Counterfactuals. Causal Learning and Reasoning. 2023.
  27. Colombo D, Maathuis MH. Order-independent constraint-based causal structure learning. Journal of Machine Learning Research. 2014;15(1):3741-82.
  28. Causation, Prediction, and Search. MIT Press; 2000.
  29. Estimating mutual information. Physical Review E. 2004;69(6):066138.
  30. Chen T, Guestrin C. XGBoost: A scalable tree boosting system. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; 2016. p. 785-94.
  31. Friedman JH. Multivariate adaptive regression splines. The Annals of Statistics. 1991;19(1):1-67.
  32. Approximate kernel-based conditional independence tests for fast non-parametric causal discovery. Journal of Causal Inference. 2019;7(1).
  33. DirectLiNGAM: A direct method for learning a linear non-Gaussian structural equation model. Journal of Machine Learning Research. 2011;12:1225-48.
  34. Hyvärinen A, Smith SM. Pairwise likelihood ratios for estimation of non-Gaussian structural equation models. Journal of Machine Learning Research. 2013;14(Jan):111-52.
  35. Distinguishing cause from effect using observational data: methods and benchmarks. Journal of Machine Learning Research. 2016;17(1):1103-204.
  36. Lasko TA, Mesa DA. Computational Phenotype Discovery via Probabilistic Independence. KDD Workshop on Applied Data Science for Healthcare. 2019.
  37. A similarity measure for indefinite rankings. ACM Transactions on Information Systems (TOIS). 2010;28(4):1-38.
  38. Fleming TR, Harrington DP. Counting Processes and Survival Analysis. vol. 169. John Wiley & Sons; 2011.
  39. Hirschfield GM, Gershwin ME. The immunobiology and pathophysiology of primary biliary cirrhosis. Annual Review of Pathology: Mechanisms of Disease. 2013;8:303-30.
  40. Bilirubin alone as a biomarker for short-term mortality in acute-on-chronic liver failure: an important prognostic indicator. Annals of Hepatology. 2014;13(1):98-104.
  41. Child CG. Surgery and portal hypertension. The Liver and Portal Hypertension. 1964:50-2.
  42. Using the ADAP learning algorithm to forecast the onset of diabetes mellitus. In: Proceedings of the Annual Symposium on Computer Application in Medical Care. American Medical Informatics Association; 1988. p. 261.
Citations (27)

Summary

We haven't generated a summary for this paper yet.