Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Causal-StoNet: Causal Inference for High-Dimensional Complex Data (2403.18994v1)

Published 27 Mar 2024 in stat.ML and cs.LG

Abstract: With the advancement of data science, the collection of increasingly complex datasets has become commonplace. In such datasets, the data dimension can be extremely high, and the underlying data generation process can be unknown and highly nonlinear. As a result, the task of making causal inference with high-dimensional complex data has become a fundamental problem in many disciplines, such as medicine, econometrics, and social science. However, the existing methods for causal inference are frequently developed under the assumption that the data dimension is low or that the underlying data generation process is linear or approximately linear. To address these challenges, this paper proposes a novel causal inference approach for dealing with high-dimensional complex data. The proposed approach is based on deep learning techniques, including sparse deep learning theory and stochastic neural networks, that have been developed in recent literature. By using these techniques, the proposed approach can address both the high dimensionality and unknown data generation process in a coherent way. Furthermore, the proposed approach can also be used when missing values are present in the datasets. Extensive numerical studies indicate that the proposed approach outperforms existing ones.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (51)
  1. Stability of stochastic approximation under verifiable conditions. SIAM Journal on Control and Optimization, 44(1):283–312, 2005.
  2. High-dimensional confounding adjustment using continuous spike and slab priors. Bayesian analysis, 14 3:805–828, 2019.
  3. Recursive partitioning for heterogeneous causal effects. Proceedings of the National Academy of Sciences, 113:7353 – 7360, 2015. URL https://api.semanticscholar.org/CorpusID:16171120.
  4. Approximate residual balancing: De-biased inference of average treatment effects in high dimensions. Journal of the Royal Statistical Society. Series B, Statistical methodology, 80(4):597–623, 2018.
  5. Inference on treatment effects after selection amongst high-dimensional controls. Review of Economic Studies, 81:608–650, 2014.
  6. Adaptive Algorithms and Stochastic Approximations. Berlin: Springer, 1990.
  7. Peter Bühlmann. Causal statistical inference in high dimensions. Mathematical Methods of Operations Research, 77:357–370, 2013.
  8. Stochastic gradient hamiltonian monte carlo. In International conference on machine learning, pp. 1683–1691, 2014.
  9. Causal inference of general treatment effects using neural networks with a diverging number of confounders. Journal of Econometrics, 238(1):105555, 2024. ISSN 0304-4076. doi: https://doi.org/10.1016/j.jeconom.2023.105555. URL https://www.sciencedirect.com/science/article/pii/S0304407623002713.
  10. Double/debiased machine learning for treatment and structural parameters. The Econometrics Journal, 21, 2018.
  11. Model-free feature screening for ultrahigh dimensional discriminant analysis. Journal of the American Statistical Association, 110(510):630–641, 2015.
  12. An adaptive empirical bayesian method for sparse deep learning. Advances in neural information processing systems, 2019:5563–5573, 2019.
  13. M. Farrell. Robust inference on average treatment effects with possibly more covariates than observations. Journal of Econometrics, 189:1–23, 2015.
  14. Deep neural networks for estimation and inference. Econometrica, 89:181–213, 2021.
  15. Global convergence of stochastic gradient hamiltonian monte carlo for nonconvex stochastic optimization: Nonasymptotic performance bounds and momentum-based acceleration. Operations Research, 2021.
  16. A unified framework for causal inference with multiple imputation using martingale. arXiv: Methodology, 2019.
  17. Estimating average treatment effects via orthogonal regularization. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, pp.  680–689, 2021.
  18. Guido Imbens. Nonparametric estimation of average treatment effects under exogeneity: A review. The Review of Economics and Statistics, 86:4–29, 2004.
  19. Wenxin Jiang. Bayesian variable selection for high dimensional generalized linear models: convergence rates of the fitted densities. The Annals of Statistics, 35(4):1487–1511, 2007.
  20. Estimating causal effects using a multi-task deep ensemble. In Proceedings of the 40 th International Conference on Machine Learning (ICML), PMLR, pp.  680–689, 2023.
  21. Metalearners for estimating heterogeneous treatment effects using machine learning. Proceedings of the National Academy of Sciences of the United States of America, 116:4156 – 4165, 2017. URL https://api.semanticscholar.org/CorpusID:73455742.
  22. Conformal Inference of Counterfactuals and Individual Treatment Effects. Journal of the Royal Statistical Society Series B: Statistical Methodology, 83(5):911–938, 10 2021. ISSN 1369-7412. doi: 10.1111/rssb.12445. URL https://doi.org/10.1111/rssb.12445.
  23. Causal decision trees. IEEE Transactions on Knowledge and Data Engineering, 29:257–271, 2015. URL https://api.semanticscholar.org/CorpusID:5718772.
  24. stochastic approximation in monte carlo computation. Journal of the American Statistical Association, 102:305–320, 2007.
  25. Bayesian subset modeling for high dimensional generalized linear models. Journal of the American Statistical Association, 108:589–606, 2013.
  26. Simulated stochastic approximation annealing for global optimization with a square-root cooling schedule. Journal of the American Statistical Association, 109:847–863, 2014.
  27. An imputation-regularized optimization algorithm for high-dimensional missing data problems and beyond. Journal of the Royal Statistical Society, Series B, 80(5):899–926, 2018a.
  28. Bayesian neural networks for selection of drug sensitive genes. Journal of the American Statistical Association, 113:955–972, 2018b.
  29. Nonlinear sufficient dimension reduction with a stochastic neural network. NeurIPS, 2022.
  30. Causal effect inference with deep latent-variable models. 31st Conference on Neural Information Processing Systems (NIPS 2017), 2017.
  31. Judea Pearl. Direct and indirect effects. In Proceedings of the 17th Conference on Uncertainty in Artificial Intelligence (UAI), pp.  411–420, 2001.
  32. Non-convex learning via stochastic gradient langevin dynamics: a nonasymptotic analysis. In Conference on Learning Theory, pp.  1674–1703. PMLR, 2017.
  33. A stochastic approximation method. The annals of mathematical statistics, pp.  400–407, 1951.
  34. Estimation of regression coefficients when some regressors are not always observed. Journal of the American Statistical Association, 89:846–866, 1994. URL https://api.semanticscholar.org/CorpusID:120769390.
  35. P. R. Rosenbaum. Observational Studies (2nd edition). Springer, New York, 2002.
  36. Donald B. Rubin. Estimating causal effects of treatments in randomized and nonrandomized studies. Journal of Educational Psychology, 66:688–701, 1974.
  37. Learning counterfactual representations for estimating individual dose-response curves. ArXiv, abs/1902.00981, 2020.
  38. Hi-ci: Deep causal inference in high dimensions. Journal of Machine Learning Research, 2020.
  39. Adapting neural networks for the estimation of treatment effects. In NeurIPS, 2019.
  40. Extended stochastic gradient mcmc for large-scale bayesian variable selection. Biometrika, 107(4):997–1004, 2020.
  41. Consistent sparse deep learning: Theory and computation. Journal of the American Statistical Association, 117(540):1981–1995, 2022.
  42. A kernel-expanded stochastic neural network. Journal of the Royal Statistical Society Series B, 84(2):547–578, 2022.
  43. Sparse deep learning: A new framework immune to local traps and miscalibration. NeurIPS 2021, 2021.
  44. R. Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society, Series B, 58:267–288, 1996.
  45. Mark J. van der Laan and Daniel Rubin. Targeted maximum likelihood learning. The International Journal of Biostatistics, 2(1):11, 2006.
  46. Super learner. Statistical Applications in Genetics and Molecular Biology, 6, 2007. URL https://api.semanticscholar.org/CorpusID:7162180.
  47. Bayesian learning via stochastic gradient Langevin dynamics. In ICML, 2011.
  48. Consistency and fluctuations for stochastic gradient langevin dynamics. Journal of Machine Learning Research, 2016.
  49. Markers associated with tumor recurrence in patients with breast cancer achieving a pathologic complete response after neoadjuvant chemotherapy. Frontiers in Oncology, 12, 2022. URL https://api.semanticscholar.org/CorpusID:248268528.
  50. Causal inference with confounders missing not at random. Biometrika, 106:875–888, 2019.
  51. Ganite: Estimation of individualized treatment effects using generative adversarial nets. In International Conference on Learning Representations, 2018. URL https://api.semanticscholar.org/CorpusID:65516833.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets