Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Integrating Uncertainty Awareness into Conformalized Quantile Regression (2306.08693v2)

Published 14 Jun 2023 in stat.ME and stat.ML

Abstract: Conformalized Quantile Regression (CQR) is a recently proposed method for constructing prediction intervals for a response $Y$ given covariates $X$, without making distributional assumptions. However, existing constructions of CQR can be ineffective for problems where the quantile regressors perform better in certain parts of the feature space than others. The reason is that the prediction intervals of CQR do not distinguish between two forms of uncertainty: first, the variability of the conditional distribution of $Y$ given $X$ (i.e., aleatoric uncertainty), and second, our uncertainty in estimating this conditional distribution (i.e., epistemic uncertainty). This can lead to intervals that are overly narrow in regions where epistemic uncertainty is high. To address this, we propose a new variant of the CQR methodology, Uncertainty-Aware CQR (UACQR), that explicitly separates these two sources of uncertainty to adjust quantile regressors differentially across the feature space. Compared to CQR, our methods enjoy the same distribution-free theoretical coverage guarantees, while demonstrating in our experiments stronger conditional coverage properties in simulated settings and real-world data sets alike.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (41)
  1. Tennessee’s student teacher achievement ratio (STAR) project. Accessed: July, 2019.
  2. Machine learning approach of automatic identification and counting of blood cells. Healthcare Technology Letters, 6(4):103–108.
  3. The limits of distribution-free conditional predictive inference. Information and Inference: A Journal of the IMA, 10(2):455–482.
  4. Breiman, L. (1996). Bagging predictors. Machine Learning, 24(2):123–140.
  5. Quantile and probability curves without crossing. Econometrica, 78(3):1093–1125.
  6. Distributional conformal prediction. Proceedings of the National Academy of Sciences, 118(48):e2107794118.
  7. Deepglobe 2018: A challenge to parse the earth through satellite images. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops.
  8. UCI machine learning repository.
  9. Efron, B. (1981). Nonparametric standard errors and confidence intervals. Canadian Journal of Statistics, 9(2):139–158.
  10. An Introduction to the Bootstrap. Number 57 in Monographs on Statistics and Applied Probability. Chapman & Hall/CRC, Boca Raton, Florida, USA.
  11. Event labeling combining ensemble detectors and background knowledge. Progress in Artificial Intelligence, pages 1–15.
  12. Improving conditional coverage via orthogonal quantile regression. Advances in neural information processing systems, 34:2060–2071.
  13. Strictly proper scoring rules, prediction, and estimation. Journal of the American Statistical Association, 102(477):359–378.
  14. Why do tree-based models still outperform deep learning on tabular data? arXiv preprint arXiv:2207.08815.
  15. Nested conformal prediction and quantile out-of-bag ensemble methods. Pattern Recognition, 127:108496.
  16. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778.
  17. Snapshot ensembles: Train 1, get m for free. In International Conference on Learning Representations.
  18. Aleatoric and epistemic uncertainty in machine learning: An introduction to concepts and methods. Machine Learning, 110:457–506.
  19. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International conference on machine learning, pages 448–456. pmlr.
  20. Flexible distribution-free conditional predictive bands using density estimators. In Chiappa, S. and Calandra, R., editors, Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics, volume 108 of Proceedings of Machine Learning Research, pages 3068–3077. PMLR.
  21. Kaggle (2016). House sales in King County, USA. https://www.kaggle.com/harlfoxem/housesalesprediction/metadata. Accessed: August, 2019.
  22. Predictive inference is free with the jackknife+-after-bootstrap. Advances in Neural Information Processing Systems, 33:4138–4149.
  23. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
  24. Adaptive, distribution-free prediction intervals for deep networks. In International Conference on Artificial Intelligence and Statistics, pages 4346–4356. PMLR.
  25. Koenker, R. (1994). Confidence intervals for regression quantiles. In Mandl, P. and Hušková, M., editors, Asymptotic Statistics, pages 349–359, Heidelberg. Physica-Verlag HD.
  26. Simple and scalable predictive uncertainty estimation using deep ensembles. Advances in neural information processing systems, 30.
  27. Distribution-free predictive inference for regression. Journal of the American Statistical Association, 113(523):1094–1111.
  28. Distribution-free prediction bands for non-parametric regression. Journal of the Royal Statistical Society: Series B: Statistical Methodology, pages 71–96.
  29. Mammen, E. (1991). Nonparametric regression under qualitative smoothness assumptions. The Annals of Statistics, 19(2):741–759.
  30. Do deep neural networks learn shallow learnable examples first? In ICML 2019 Workshop on Identifying and Understanding Deep Learning Phenomena.
  31. Meinshausen, N. (2006). Quantile regression forests. Journal of Machine Learning Research, 7(35):983–999.
  32. Neal, R. M. (2012). Bayesian learning for neural networks, volume 118. Springer Science & Business Media.
  33. A data-driven software tool for enabling cooperative information sharing among police departments. European Journal of Operational Research, 141(3):660–678.
  34. Conformalized quantile regression. In NeurIPS.
  35. Dex: Deep expectation of apparent age from a single image. In IEEE International Conference on Computer Vision Workshops (ICCVW).
  36. Deep expectation of real and apparent age from a single image without facial landmarks. International Journal of Computer Vision, 126(2-4):144–157.
  37. Reliable classification: Learning classifiers that distinguish aleatoric and epistemic uncertainty. Information Sciences, 255:16–29.
  38. A comparison of some conformal quantile regression methods. Stat, 9(1):e261.
  39. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research, 15(1):1929–1958.
  40. Algorithmic Learning in a Random World. Springer-Verlag, Berlin, Heidelberg.
  41. Yeh, I.-C. (1998). Modeling of strength of high-performance concrete using artificial neural networks. Cement and Concrete research, 28(12):1797–1808.
Citations (4)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets