Evaluating the Effectiveness of Index-Based Treatment Allocation (2402.11771v1)
Abstract: When resources are scarce, an allocation policy is needed to decide who receives a resource. This problem occurs, for instance, when allocating scarce medical resources and is often solved using modern ML methods. This paper introduces methods to evaluate index-based allocation policies -- that allocate a fixed number of resources to those who need them the most -- by using data from a randomized control trial. Such policies create dependencies between agents, which render the assumptions behind standard statistical tests invalid and limit the effectiveness of estimators. Addressing these challenges, we translate and extend recent ideas from the statistics literature to present an efficient estimator and methods for computing asymptotically correct confidence intervals. This enables us to effectively draw valid statistical conclusions, a critical gap in previous work. Our extensive experiments validate our methodology in practical settings, while also showcasing its statistical power. We conclude by proposing and empirically verifying extensions of our methodology that enable us to reevaluate a past randomized control trial to evaluate different ML allocation policies in the context of a mHealth program, drawing previously invisible conclusions.
- Tom M Apostol. 1974. Mathematical Analysis. Addison-Wesley.
- Susan Athey and Stefan Wager. 2021. Policy learning with observational data. Econometrica 89, 1 (2021), 133–161.
- Prioritizing hepatitis C treatment in US prisons. Operations Research 67, 3 (2019), 853–873.
- Efficient and targeted COVID-19 border testing via reinforcement learning. Nature 599, 7883 (2021), 108–113.
- Debopam Bhattacharya and Pascaline Dupas. 2012. Inferring welfare maximizing treatment assignment under budget constraints. Journal of Econometrics 167, 1 (2012), 168–196.
- Philip E Cheng. 1984. Strong consistency of nearest neighbor regression function estimators. Journal of Multivariate Analysis 15, 1 (1984), 63–72.
- Limited Resource Allocation in a Non-Markovian World: The Case of Maternal and Child Healthcare. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence (IJCAI ’23). ijcai.org, 5950–5958.
- Improving health outcomes through better capacity allocation in a community-based chronic care model. Operations Research 61, 6 (2013), 1277–1294.
- Planning for HIV screening, testing, and care at the veterans health administration. Operations research 63, 2 (2015), 287–304.
- Luc Devroye. 1978. The uniform convergence of nearest neighbor regression function estimators and their application in optimization. IEEE Transactions on Information Theory 24, 2 (1978), 142–151.
- Scaling Multi-Armed Bandit Algorithms. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD ’19). ACM, 1449–1459.
- Data-driven predictive maintenance scheduling policies for railways. Transportation Research Part C: Emerging Technologies 107 (2019), 137–154.
- Eduardo Hariton and Joseph J Locascio. 2018. Randomised controlled trials—the gold standard for effectiveness research. BJOG: An International Journal of Obstetrics and Gynaecology 125, 13 (2018), 1716.
- Kosuke Imai and Michael Lingzhi Li. 2023. Statistical Inference for Heterogeneous Treatment Effects Discovered by Generic Machine Learning in Randomized Experiments. arXiv:2203.14511v2 [stat.ME]
- Guido W Imbens and Donald B Rubin. 2015. Causal inference in statistics, social, and biomedical sciences. Cambridge University Press.
- The risks and rewards of covariate adjustment in randomized trials: an assessment of 12 outcomes from 8 studies. Trials 15, 1 (2014), 1–7.
- Edward H Kennedy. 2023. Towards optimal doubly robust estimation of heterogeneous causal effects. Electronic Journal of Statistics 17, 2 (2023), 3008–3049.
- Risk and treatment effect heterogeneity: re-analysis of individual participant data from 32 large clinical trials. International journal of epidemiology 45, 6 (2016), 2075–2088.
- Q-Learning Lagrange Policies for Multi-Action Restless Bandits. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD ’21). ACM, 871–881.
- Equitable Restless Multi-Armed Bandits: A General Framework Inspired By Digital Health. arXiv:2308.09726 [cs.LG]
- Learning to prescribe interventions for tuberculosis patients using digital adherence data. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD ’19). ACM, 2430–2438.
- Toru Kitagawa and Aleksey Tetenov. 2018. Who should be treated? empirical welfare maximization methods for treatment choice. Econometrica 86, 2 (2018), 591–616.
- Metalearners for estimating heterogeneous treatment effects using machine learning. Proceedings of the National Academy of Sciences 116, 10 (2019), 4156–4165.
- A model for allocating CDC’s HIV prevention resources in the United States. Health Care Management Science 14 (2011), 115–124.
- Optimal screening for hepatocellular carcinoma: A restless bandit model. Manufacturing & Service Operations Management 21, 1 (2019), 198–212.
- Alexander R Luedtke and Mark J van der Laan. 2016. Optimal individualized treatments in resource-limited settings. The International Journal of Biostatistics 12, 1 (2016), 283–303.
- Jesus Luque and Daniel Straub. 2019. Risk-based optimal inspection strategies for structural systems using dynamic Bayesian networks. Structural Safety 76 (2019), 68–80.
- An efficacy study of a ninth-grade early warning indicator intervention. Journal of Research on Educational Effectiveness 12, 3 (2019), 363–390.
- Collapsing Bandits and Their Application to Public Health Intervention. In Proceedings of the Thirty-fourth Annual Conference on Neural Information Processing Systems (NeurIPs ’20).
- Field study in deploying restless multi-armed bandits: Assisting non-profits in improving maternal and child health. In Proceedings of the 36th AAAI Conference on Artificial Intelligence (AAAI ’22). 12017–12025.
- Improved Policy Evaluation for Randomized Trials of Algorithmic Resource Allocation. In Proceedings of the 40th International Conference on Machine Learning (ICML ’23). 24198–24213.
- Difficult Lessons on Social Prediction from Wisconsin Public Schools. arXiv:2304.06205 [cs.CY]
- Bodhisattva Sen. 2018. A gentle introduction to empirical process theory and applications. Lecture Notes, Columbia University 11 (2018), 28–29.
- Stephen S Senn. 2008. Statistical issues in drug development. Vol. 69. John Wiley & Sons.
- Treatment allocation under uncertain costs. arXiv:2103.11066 [stat.ME]
- Liyang Sun. 2021. Empirical welfare maximization with constraints. arXiv:2103.15298 [econ.EM]
- Milind Tambe. 2022. AI for Social Impact: Results from Deployments for Public Health and Conversation. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD ’22). ACM, 2.
- Aad van der Vaart. 2000. Asymptotic statistics. Vol. 3. Cambridge university press.
- Aad van der Vaart and Jon A Wellner. 2023. Empirical processes. In Weak Convergence and Empirical Processes: With Applications to Statistics. Springer, 127–384.
- Restless Multi-Armed Bandits for Maternal and Child Health: Results from Decision-Focused Learning. In Proceedings of the 22nd International Conference on Autonomous Agents and Multiagent Systems (AAMAS ’23). 1312–1320.
- Expanding impact of mobile health programs: SAHELI for maternal and child care. AI Magazine 44, 4 (2023), 363–376.
- Stefan Wager and Susan Athey. 2018. Estimation and inference of heterogeneous treatment effects using random forests. Journal of the American Statistical Association 113, 523 (2018), 1228–1242.
- Scalable decision-focused learning in restless multi-armed bandits with application to maternal and child health. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 37. 12138–12146.
- Richard R Weber and Gideon Weiss. 1990. On an index policy for restless bandits. Journal of Applied Probability 27, 3 (1990), 637–648.
- Jeffrey M. Wooldridge. 2019. Introductory Econometrics: A Modern Approach. Cengage Learning. Chapter 7.
- Evaluating treatment prioritization rules via rank-weighted average treatment effects. arXiv:2111.07966 [stat.ME]
- Risk-based maintenance planning of offshore wind turbine farms. Reliability Engineering & System Safety 202 (2020), 107062.
- Efficient augmentation and relaxation learning for individualized treatment rules using observational data. Journal of Machine Learning Research 20 (2019), 48:1–48:23.
- Estimating individualized treatment rules using outcome weighted learning. Journal of the American Statistical Association 107, 499 (2012), 1106–1118.