Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Causal Effect Estimation Using Random Hyperplane Tessellations (2404.10907v3)

Published 16 Apr 2024 in cs.AI

Abstract: Matching is one of the simplest approaches for estimating causal effects from observational data. Matching techniques compare the observed outcomes across pairs of individuals with similar covariate values but different treatment statuses in order to estimate causal effects. However, traditional matching techniques are unreliable given high-dimensional covariates due to the infamous curse of dimensionality. To overcome this challenge, we propose a simple, fast, yet highly effective approach to matching using Random Hyperplane Tessellations (RHPT). First, we prove that the RHPT representation is an approximate balancing score -- thus maintaining the strong ignorability assumption -- and provide empirical evidence for this claim. Second, we report results of extensive experiments showing that matching using RHPT outperforms traditional matching techniques and is competitive with state-of-the-art deep learning methods for causal effect estimation. In addition, RHPT avoids the need for computationally expensive training of deep neural networks.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. On the surprising behavior of distance metrics in high dimensional spaces. In International Conference on Database Theory, 2001.
  2. When is ”nearest neighbor” meaningful? In International Conference on Database Theory, 1999.
  3. Moses S. Charikar. Similarity estimation techniques from rounding algorithms. In Proceedings of the Thiry-Fourth Annual ACM Symposium on Theory of Computing, STOC ’02, page 380–388, New York, NY, USA, 2002. Association for Computing Machinery. ISBN 1581134959. doi:10.1145/509907.509965.
  4. David Roxbee Cox. Planning of experiments. 1958.
  5. Sharp estimates on random hyperplane tessellations. SIAM Journal on Mathematics of Data Science, 4(4):1396–1419, 2022. doi:10.1137/22M1485826.
  6. Locality preserving projections. In Advances in Neural Information Processing Systems, volume 16. MIT Press, 2003.
  7. Causal Inference - What if. Taylor and Francis, 2023.
  8. Jennifer L. Hill. Bayesian nonparametric modeling for causal inference. Journal of Computational and Graphical Statistics, 20:217 – 240, 2011.
  9. Approximate nearest neighbors: Towards removing the curse of dimensionality. In Proceedings of the Thirtieth Annual ACM Symposium on Theory of Computing, STOC ’98, page 604–613. Association for Computing Machinery, 1998.
  10. Quantifying ignorance in individual-level causal-effect estimates under hidden confounding. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 4829–4838. PMLR, 18–24 Jul 2021.
  11. Learning representations for counterfactual inference. In International Conference on Machine Learning, 2016.
  12. Why propensity scores should not be used for matching. Political analysis, 27(4):435–454, 2019.
  13. Adam: A method for stochastic optimization. CoRR, abs/1412.6980, 2014.
  14. Deep learning for causal inference, Oct 2021. URL osf.io/preprints/socarxiv/aeszf.
  15. Yann LeCun. The mnist database of handwritten digits, 1998. URL http://yann.lecun.com/exdb/mnist/.
  16. Matching via dimensionality reduction for estimation of treatment effects in digital marketing campaigns. IJCAI’16, page 3768–3774, 2016.
  17. Causal effect inference with deep latent-variable models. Advances in neural information processing systems, 30, 2017.
  18. Yurii Nesterov. A method for solving the convex programming problem with convergence rate 𝒪⁢(1/k2)𝒪1superscript𝑘2\mathcal{O}(1/k^{2})caligraphic_O ( 1 / italic_k start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ). Proceedings of the USSR Academy of Sciences, 269:543–547, 1983.
  19. David Newman. Bag of words data set, 2008. URL https://archive.ics.uci.edu/ml/datasets/Bag+of+Words.
  20. Dimension reduction by random hyperplane tessellations. Discrete Comput. Geom., 51(2):438–461, mar 2014. doi:10.1007/s00454-013-9561-6.
  21. Iacopo Poli. Article on efficient neural networks training through locality sensitive hashing. medium.com, 2021.
  22. Scaling up simhash. In Sinno Jialin Pan and Masashi Sugiyama, editors, Proceedings of The 12th Asian Conference on Machine Learning, volume 129 of Proceedings of Machine Learning Research, pages 705–720. PMLR, 18–20 Nov 2020.
  23. The central role of the propensity score in observational studies for causal effects. Biometrika, 70(1):41–55, 04 1983. ISSN 0006-3444. doi:10.1093/biomet/70.1.41.
  24. Donald B. Rubin. Bias reduction using mahalanobis-metric matching. Biometrics, 36(2):293–298, 1980.
  25. Estimating individual treatment effect: generalization bounds and algorithms. In International Conference on Machine Learning, pages 3076–3085. PMLR, 2017.
  26. Adapting neural networks for the estimation of treatment effects. Advances in neural information processing systems, 32, 2019.
  27. Elizabeth A. Stuart. Matching methods for causal inference: A review and a look forward. Statistical science : a review journal of the Institute of Mathematical Statistics, 25 1:1–21, 2010.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets