Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Statistical Performance Guarantee for Subgroup Identification with Generic Machine Learning (2310.07973v2)

Published 12 Oct 2023 in stat.ME, math.OC, stat.AP, and stat.ML

Abstract: Across a wide array of disciplines, many researchers use ML algorithms to identify a subgroup of individuals who are likely to benefit from a treatment the most (``exceptional responders'') or those who are harmed by it. A common approach to this subgroup identification problem consists of two steps. First, researchers estimate the conditional average treatment effect (CATE) using an ML algorithm. Next, they use the estimated CATE to select those individuals who are predicted to be most affected by the treatment, either positively or negatively. Unfortunately, CATE estimates are often biased and noisy. In addition, utilizing the same data to both identify a subgroup and estimate its group average treatment effect results in a multiple testing problem. To address these challenges, we develop uniform confidence bands for estimation of the group average treatment effect sorted by generic ML algorithm (GATES). Using these uniform confidence bands, researchers can identify, with a statistical guarantee, a subgroup whose GATES exceeds a certain effect size, regardless of how this effect size is chosen. The validity of the proposed methodology depends solely on randomization of treatment and random sampling of units. Importantly, our method does not require modeling assumptions and avoids a computationally intensive resampling procedure. A simulation study shows that the proposed uniform confidence bands are reasonably informative and have an appropriate empirical coverage even when the sample size is as small as 100. We analyze a clinical trial of late-stage prostate cancer and find a relatively large proportion of exceptional responders.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. Subgroup identification in clinical trials via the predicted individual treatment effect. PloS one 13, 10, e0205971.
  2. Identifying exceptional responders in randomized trials: An optimization approach. Informs Journal on Optimization 1, 3, 187–199.
  3. Bhattacharya, P. (1974). Convergence of sample paths of normalized sums of induced order statistics. Ann. Statist. 2, 1, 1034–1039.
  4. Bivariate dependence properties of order statistics. journal of multivariate analysis 56, 1, 75–89.
  5. A graphical method to assess treatment–covariate interactions using the cox model on subsets of the data. Statistics in medicine 19, 19, 2595–2609.
  6. On the rate of convergence in invariance principle. Lecture notes in mathematics 1021, 59–66.
  7. The choice of treatment for cancer patients based on covariate information. Bulletin du cancer 67, 4, 477–490.
  8. Generic machine learning inference on heterogeneous treatment effects in randomized experiments. Tech. rep., arXiv:1712.04802.
  9. Quantile and probability curves without crossing. Econometrica 78, 3, 1093–1125.
  10. Bart: Bayesian additive regression trees. The Annals of Applied Statistics 4, 1, 266–298.
  11. Quantile-based subgroup identification for randomized clinical trials. Statistics in Biosciences 13, 90–128.
  12. The Exceptional Responders Initiative: Feasibility of a National Cancer Institute Pilot Study. JNCI: Journal of the National Cancer Institute 113, 1, 27–37.
  13. Cuadras, C. M. (2002). On the covariance between functions. Journal of Multivariate Analysis 81, 1, 19–27.
  14. The asymptotic theory of concomitants of order statistics. Journal of Applied Probability 11, 4, 762–770.
  15. Functional limit theorems for induced order statistics. Mathematical Methods of Statistics 9, 3, 297–313.
  16. Automated versus do-it-yourself methods for causal inference: Lessons learned from a data analysis competition. Statistical Science 34, 1, 43–68.
  17. Subgroup identification from randomized clinical trial data. Statistics in medicine 30, 24, 2867–2880.
  18. Bayesian regression tree models for causal inference: regularization, confounding, and heterogeneous effects. Bayesian Analysis 15, 3, 965–1056.
  19. Understanding heterogeneity in response to antidiabetes treatment: a post hoc analysis using sides, a subgroup identification algorithm. Journal of Diabetes Science and Technology 7, 2, 420–430.
  20. Hill, J. L. (2011). Bayesian nonparametric modeling for causal inference. Journal of Computational and Graphical Statistics 20, 1, 217–240.
  21. Statistical inference for heterogeneous treatment effects discovered by generic machine learning in randomized experiments. arXiv preprint https://arxiv.org/pdf/2203.14511.pdf.
  22. Experimental evaluation of individualized treatment rules. Journal of the American Statistical Association 118, 541, 242–256.
  23. Estimating treatment effect heterogeneity in randomized program evaluation. Annals of Applied Statistics 7, 1, 443–470.
  24. Kahane, J.-P. (1985). Le chaos multiplicatif. Comptes rendus de l’Académie des sciences. Série 1, Mathématique 301, 6, 329–332.
  25. Responder identification in clinical trials with censored data. Computational Statistics & Data Analysis 50, 5, 1338–1355.
  26. Confidence bands for brownian motion and applications to monte carlo simulation. Statistics and Computing 17, 1–10.
  27. Neyman, J. (1923). On the application of probability theory to agricultural experiments: Essay on principles, section 9. (translated in 1990). Statistical Science 5, 465–480.
  28. Ossiander, M. (1987). A central limit theorem under metric entropy with l2 bracketing. The Annals of Probability 897–919.
  29. Radcliffe, N. J. (2007). Using control groups to target on predicted lift: Building and assessing uplift models. Direct Marketing Analytics Journal 1, 3, 14–21.
  30. Rosenkranz, G. K. (2016). Exploratory subgroup analysis in clinical trials by model selection. Biometrical Journal 58, 5, 1217–1228.
  31. Rubin, D. B. (1990). Comments on “On the application of probability theory to agricultural experiments. Essay on principles. Section 9” by J. Splawa-Neyman translated from the Polish and edited by D. M. Dabrowska and T. P. Speed. Statistical Science 5, 472–480.
  32. Sen, P. K. (1976). A note on invariance principles for induced order statistics. The Annals of Probability 4, 3, 474–479.
  33. Slepian, D. (1962). The one-sided barrier problem for gaussian noise. Bell System Technical Journal 41, 2, 463–501.
  34. Generalized gaussian bridges. Stochastic Processes and their Applications 124, 9, 3084–3105.
  35. Subgroup analysis via recursive partitioning. Journal of Machine Learning Research 10, 2.
  36. Exceptional responders—discovering predictive biomarkers. Nature Reviews Clinical Oncology 12, 3, 132–134.
  37. Tibshirani, R. (1996). Regression shrinkage and selection via LASSO. Journal of the Royal Statistical Society, Series B (Statistical Methodology) 58, 1, 267–288.
  38. Estimation and inference of heterogeneous treatment effects using random forests. Journal of the American Statistical Association 113, 523, 1228–1242.
  39. Evaluating treatment prioritization rules via rank-weighted average treatment effects. arXiv preprint 2111.07966 .
  40. Yang, S. (1977). General distribution theory of the concomitants of order statistics. The Annals of Statistics 5, 5, 996–1002.

Summary

We haven't generated a summary for this paper yet.