Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Stochastic Online Conformal Prediction with Semi-Bandit Feedback (2405.13268v2)

Published 22 May 2024 in cs.LG

Abstract: Conformal prediction has emerged as an effective strategy for uncertainty quantification by modifying a model to output sets of labels instead of a single label. These prediction sets come with the guarantee that they contain the true label with high probability. However, conformal prediction typically requires a large calibration dataset of i.i.d. examples. We consider the online learning setting, where examples arrive over time, and the goal is to construct prediction sets dynamically. Departing from existing work, we assume semi-bandit feedback, where we only observe the true label if it is contained in the prediction set. For instance, consider calibrating a document retrieval model to a new domain; in this setting, a user would only be able to provide the true label if the target document is in the prediction set of retrieved documents. We propose a novel conformal prediction algorithm targeted at this setting, and prove that it obtains sublinear regret compared to the optimal conformal predictor. We evaluate our algorithm on a retrieval task, an image classification task, and an auction price-setting task, and demonstrate that it empirically achieves good performance compared to several baselines.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (15)
  1. Practical adversarial multivalid conformal prediction. Advances in Neural Information Processing Systems 35, 29362–29373.
  2. Distribution-free, risk-controlling prediction sets. Journal of the ACM (JACM) 68(6), 1–34.
  3. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp.  248–255. Ieee.
  4. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929.
  5. Adaptive conformal inference under distribution shift. Advances in Neural Information Processing Systems 34, 1660–1672.
  6. Conformal inference for online prediction with arbitrary distribution shifts. arXiv preprint arXiv:2208.08401.
  7. Survey of hallucination in natural language generation. ACM Computing Surveys 55(12), 1–38.
  8. Dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2004.04906.
  9. Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in Neural Information Processing Systems 33, 9459–9474.
  10. Massart, P. (1990). The tight constant in the dvoretzky-kiefer-wolfowitz inequality. The annals of Probability, 1269–1283.
  11. Pac confidence sets for deep neural networks via calibrated prediction. arXiv preprint arXiv:2001.00106.
  12. Squad: 100,000+ questions for machine comprehension of text. arXiv preprint arXiv:1606.05250.
  13. Retrieval augmentation reduces hallucination in conversation. arXiv preprint arXiv:2104.07567.
  14. Conformal prediction under covariate shift. Advances in neural information processing systems 32.
  15. Algorithmic learning in a random world. Springer Science & Business Media.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Haosen Ge (4 papers)
  2. Hamsa Bastani (18 papers)
  3. Osbert Bastani (97 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.