Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Does confidence calibration improve conformal prediction? (2402.04344v3)

Published 6 Feb 2024 in cs.LG

Abstract: Conformal prediction is an emerging technique for uncertainty quantification that constructs prediction sets guaranteed to contain the true label with a predefined probability. Previous works often employ temperature scaling to calibrate classifiers, assuming that confidence calibration benefits conformal prediction. However, the specific impact of confidence calibration on conformal prediction remains underexplored. In this work, we make two key discoveries about the impact of confidence calibration methods on adaptive conformal prediction. Firstly, we empirically show that current confidence calibration methods (e.g., temperature scaling) typically lead to larger prediction sets in adaptive conformal prediction. Secondly, by investigating the role of temperature value, we observe that high-confidence predictions can enhance the efficiency of adaptive conformal prediction. Theoretically, we prove that predictions with higher confidence result in smaller prediction sets on expectation. This finding implies that the rescaling parameters in these calibration methods, when optimized with cross-entropy loss, might counteract the goal of generating efficient prediction sets. To address this issue, we propose Conformal Temperature Scaling (ConfTS), a variant of temperature scaling with a novel loss function designed to enhance the efficiency of prediction sets. This approach can be extended to optimize the parameters of other post-hoc methods of confidence calibration. Extensive experiments demonstrate that our method improves existing adaptive conformal prediction methods in classification tasks, especially with LLMs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. A gentle introduction to conformal prediction and distribution-free uncertainty quantification. arXiv preprint arXiv:2107.07511, 2021.
  2. Learn then test: Calibrating predictive algorithms to achieve risk control. arXiv preprint arXiv:2110.01052, 2021a.
  3. Uncertainty sets for image classifiers using conformal prediction. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net, 2021b.
  4. Conformal prediction for reliable machine learning: theory, adaptations and applications. Newnes, 2014.
  5. Testing for outliers with conformal p-values. The Annals of Statistics, 51(1):149–178, 2023.
  6. Learning prediction intervals for regression: Generalization and calibration. In International Conference on Artificial Intelligence and Statistics, pp.  820–828. PMLR, 2021.
  7. Training conformal predictors. In Conformal and Probabilistic Prediction and Applications, pp.  55–64. PMLR, 2020.
  8. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on Computer Vision and Pattern Recognition, pp.  248–255. Ieee, 2009.
  9. An image is worth 16x16 words: Transformers for image recognition at scale. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net, 2021.
  10. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In International Conference on Machine Learning, pp.  1050–1059. PMLR, 2016.
  11. Adversarially robust conformal prediction. In International Conference on Learning Representations, 2021.
  12. Improving uncertainty quantification of deep classifiers via neighborhood conformal prediction: Novel algorithm and theoretical analysis. arXiv preprint arXiv:2303.10694, 2023.
  13. Conformal prediction with conditional guarantees. arXiv preprint arXiv:2305.12616, 2023.
  14. Semi-supervised learning by entropy minimization. Advances in Neural Information Processing Systems, 17, 2004.
  15. Prediction and outlier detection in classification problems. Journal of the Royal Statistical Society Series B: Statistical Methodology, 84(2):524–546, 2022.
  16. On calibration of modern neural networks. In International Conference on Machine Learning, pp.  1321–1330. PMLR, 2017.
  17. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.  770–778, 2016.
  18. Densely connected convolutional networks. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp.  4700–4708, 2017.
  19. Learning multiple layers of features from tiny images. 2009.
  20. Conformal prediction with large language models for multi-choice question answering. CoRR, abs/2305.18404, 2023. doi: 10.48550/ARXIV.2305.18404.
  21. Lei, J. Classification with confidence. Biometrika, 101(4):755–769, 2014.
  22. Distribution-free prediction bands for non-parametric regression. Journal of the Royal Statistical Society Series B: Statistical Methodology, 76(1):71–96, 2014.
  23. Integrative conformal p-values for powerful out-of-distribution testing with labeled outliers. arXiv preprint arXiv:2208.11111, 2022.
  24. Estimating test performance for AI medical devices under distribution shift with conformal prediction. CoRR, abs/2207.05796, 2022. doi: 10.48550/ARXIV.2207.05796.
  25. Federated conformal predictors for distributed uncertainty quantification. In International Conference on Machine Learning, pp.  22942–22964. PMLR, 2023.
  26. Calibrating deep neural networks using focal loss. Advances in Neural Information Processing Systems, 33:15288–15299, 2020.
  27. When does label smoothing help? Advances in Neural Information Processing Systems, 32, 2019.
  28. Inductive confidence machines for regression. In Machine Learning: ECML 2002: 13th European Conference on Machine Learning Helsinki, Finland, August 19–23, 2002 Proceedings 13, pp.  345–356. Springer, 2002.
  29. Regularizing neural networks by penalizing confident output distributions. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Workshop Track Proceedings. OpenReview.net, 2017.
  30. Platt, J. et al. Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. Advances in Large Margin Classifiers, 10(3):61–74, 1999.
  31. Distribution-free uncertainty quantification for classification under label shift. In Uncertainty in Artificial Intelligence, pp.  844–853. PMLR, 2021.
  32. Do imagenet classifiers generalize to imagenet? In International Conference on Machine Learning, pp.  5389–5400. PMLR, 2019.
  33. Robots that ask for help: Uncertainty alignment for large language model planners. In 7th Annual Conference on Robot Learning, 2023.
  34. Conformalized quantile regression. Advances in Neural Information Processing Systems, 32, 2019.
  35. Classification with valid and adaptive coverage. Advances in Neural Information Processing Systems, 33:3581–3591, 2020.
  36. Least ambiguous set-valued classifiers with bounded error levels. Journal of the American Statistical Association, 114(525):223–234, 2019.
  37. Improving adaptive conformal prediction using self-supervised learning. In International Conference on Artificial Intelligence and Statistics, pp.  10160–10177. PMLR, 2023.
  38. A tutorial on conformal prediction. Journal of Machine Learning Research, 9(3), 2008.
  39. Very deep convolutional networks for large-scale image recognition. In Bengio, Y. and LeCun, Y. (eds.), 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015.
  40. Smith, R. C. Uncertainty quantification: theory, implementation, and applications, volume 12. Siam, 2013.
  41. Learning optimal conformal classifiers. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net, 2022.
  42. Conformal prediction under covariate shift. Advances in Neural Information Processing Systems, 32, 2019.
  43. Machine-learning applications of algorithmic randomness. 1999.
  44. Algorithmic learning in a random world, volume 29. Springer, 2005.
  45. Wang, C. Calibration in deep learning: A survey of the state-of-the-art. arXiv preprint arXiv:2308.01222, 2023.
  46. Rethinking calibration of deep neural networks: Do not be afraid of overconfidence. Advances in Neural Information Processing Systems, 34:11809–11820, 2021.
  47. Mitigating neural network overconfidence with logit normalization. In International Conference on Machine Learning, pp.  23631–23644. PMLR, 2022.
  48. Beyond confidence: Reliable models should also consider atypicality. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  49. Obtaining calibrated probability estimates from decision trees and naive bayesian classifiers. In International Conference on Machine Learning, volume 1, pp.  609–616, 2001.
  50. Transforming classifier scores into accurate multiclass probability estimates. In Proceedings of the eighth ACM SIGKDD International Conference on Knowledge Discovery and Data mining, pp.  694–699, 2002.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Huajun Xi (4 papers)
  2. Jianguo Huang (30 papers)
  3. Lei Feng (190 papers)
  4. Hongxin Wei (45 papers)
  5. Kangdao Liu (7 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets