Not all distributional shifts are equal: Fine-grained robust conformal inference
Abstract: We introduce a fine-grained framework for uncertainty quantification of predictive models under distributional shifts. This framework distinguishes the shift in covariate distributions from that in the conditional relationship between the outcome ($Y$) and the covariates ($X$). We propose to reweight the training samples to adjust for an identifiable covariate shift while protecting against worst-case conditional distribution shift bounded in an $f$-divergence ball. Based on ideas from conformal inference and distributionally robust learning, we present an algorithm that outputs (approximately) valid and efficient prediction intervals in the presence of distributional shifts. As a use case, we apply the framework to sensitivity analysis of individual treatment effects with hidden confounding. The proposed methods are evaluated in simulation studies and four real data applications, demonstrating superior robustness and efficiency compared with existing benchmarks.
- Conformal prediction: A gentle introduction. Foundations and Trends® in Machine Learning, 16(4):494–591.
- The limits of distribution-free conditional predictive inference. Information and Inference: A Journal of the IMA, 10(2):455–482.
- Conformal prediction beyond exchangeability. The Annals of Statistics, 51(2):816–845.
- Quantifying distributional model risk via optimal transport. Mathematics of Operations Research, 44(2):565–600.
- Breiman, L. (2001). Random forests. Machine learning, 45:5–32.
- Assessing treatment effect variation in observational studies: Results from a data challenge.
- Robust validation: Confident predictions even when distributions shift. Journal of the American Statistical Association, (just-accepted):1–22.
- Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, pages 785–794.
- Distributional conformal prediction. Proceedings of the National Academy of Sciences, 118(48):e2107794118.
- Cover, T. M. (1999). Elements of information theory. John Wiley & Sons.
- Retiring adult: New datasets for fair machine learning. Advances in neural information processing systems, 34:6478–6490.
- Distributionally robust losses for latent covariate mixtures. Operations Research, 71(2):649–664.
- Learning models with uniform performance via distributionally robust optimization. The Annals of Statistics, 49(3):1378–1406.
- Elie-Dit-Cosaque, K. (2020). qosa-indices.
- Adversarially robust conformal prediction. In International Conference on Learning Representations.
- Probabilistically robust conformal prediction. In Uncertainty in Artificial Intelligence, pages 681–690. PMLR.
- Guan, L. (2023). Localized conformal prediction: A generalized inference framework for conformal prediction. Biometrika, 110(1):33–50.
- Conformalized survival analysis with adaptive cutoffs. Biometrika, page asad076.
- Nested conformal prediction and quantile out-of-bag ensemble methods. Pattern Recognition, 127:108496.
- Hastie, T. J. (2017). Generalized additive models. In Statistical models in S, pages 249–307. Routledge.
- Causal inference in statistics, social, and biomedical sciences. Cambridge University Press.
- Diagnosing the role of observable distribution shift in scientific replications. arXiv preprint arXiv:2309.01056.
- Sensitivity analysis of individual treatment effects: A robust conformal inference approach. Proceedings of the National Academy of Sciences, 120(6):e2214889120.
- Sensitivity analysis under the f𝑓fitalic_f-sensitivity models: a distributional robustness perspective. arXiv preprint arXiv:2203.04373.
- Confounding-robust policy evaluation in infinite-horizon reinforcement learning. Advances in neural information processing systems, 33:22293–22304.
- Minimax-optimal policy learning under unobserved confounding. Management Science, 67(5):2870–2890.
- Distribution-free predictive inference for regression. Journal of the American Statistical Association, 113(523):1094–1111.
- Conformal inference of counterfactuals and individual treatment effects. Journal of the Royal Statistical Society Series B: Statistical Methodology, 83(5):911–938.
- Policy learning under biased sample selection. arXiv preprint arXiv:2304.11735.
- On the need for a language describing distribution shifts: Illustrations on tabular datasets. arXiv preprint arXiv:2307.05284.
- The effect of natural distribution shift on question answering models. In International conference on machine learning, pages 6905–6916. PMLR.
- Factored DRO: Factored distributionally robust policies for contextual bandits. Advances in Neural Information Processing Systems, 35:8318–8331.
- Diagnosing model performance under distribution shift. arXiv preprint arXiv:2303.02011.
- Neyman, J. (1923). Sur les applications de la théorie des probabilités aux experiences agricoles: Essai des principes. Roczniki Nauk Rolniczych, 10(1):1–51.
- Inductive confidence machines for regression. In Machine Learning: ECML 2002: 13th European Conference on Machine Learning Helsinki, Finland, August 19–23, 2002 Proceedings 13, pages 345–356. Springer.
- PAC prediction sets under covariate shift. In International Conference on Learning Representations.
- Scikit-learn: Machine learning in python. Journal of machine learning research, 12(Oct):2825–2830.
- Fighting covid-19 misinformation on social media: Experimental evidence for a scalable accuracy-nudge intervention. Psychological science, 31(7):770–780.
- Distribution-free uncertainty quantification for classification under label shift. In Uncertainty in Artificial Intelligence, pages 844–853. PMLR.
- Distribution-free prediction sets adaptive to unknown covariate shift. arXiv preprint arXiv:2203.06126.
- Distributionally robust optimization: A review. arXiv preprint arXiv:1908.05659.
- Do imagenet classifiers generalize to imagenet? In International conference on machine learning, pages 5389–5400. PMLR.
- Conformalized quantile regression. Advances in neural information processing systems, 32.
- Classification with valid and adaptive coverage.
- How accurate are accuracy-nudge interventions? a preregistered direct replication of pennycook et al.(2020). Psychological science, 32(7):1169–1178.
- Rosenbaum, P. R. (1987). Sensitivity analysis for certain permutation inferences in matched observational studies. Biometrika, 74(1):13–26.
- Learning from a biased sample. arXiv preprint arXiv:2209.01754.
- Distributionally robust logistic regression. Advances in Neural Information Processing Systems, 28.
- Distributionally robust batch contextual bandits. Management Science.
- Tan, Z. (2006). A distributional approach for causal inference using propensity scores. Journal of the American Statistical Association, 101(476):1619–1637.
- Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society Series B: Statistical Methodology, 58(1):267–288.
- Conformal prediction under covariate shift. Advances in neural information processing systems, 32.
- Algorithmic learning in a random world, volume 29. Springer.
- External validation of a widely implemented proprietary sepsis prediction model in hospitalized patients. JAMA Internal Medicine, 181(8):1065–1070.
- Bounds on the conditional and average treatment effect with unobserved confounding factors. arXiv preprint arXiv:1808.09521.
- Doubly robust calibration of prediction sets under covariate shift. arXiv preprint arXiv:2203.01761.
- Yeager, D. S. (2019). The National Study of Learning Mindsets,[United States], 2015-2016.
- A national experiment reveals where a growth mindset improves achievement. Nature, 573(7774):364–369.
- Conformal sensitivity analysis for individual treatment effects. Journal of the American Statistical Association, pages 1–14.
- Optimal multi-distribution learning. arXiv preprint arXiv:2312.05134.
- Sensitivity analysis for inverse probability weighting estimators via the percentile bootstrap. Journal of the Royal Statistical Society Series B: Statistical Methodology, 81(4):735–761.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.