Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 93 tok/s
Gemini 2.5 Pro 56 tok/s Pro
GPT-5 Medium 25 tok/s
GPT-5 High 22 tok/s Pro
GPT-4o 98 tok/s
GPT OSS 120B 452 tok/s Pro
Kimi K2 212 tok/s Pro
2000 character limit reached

Online conformal prediction with decaying step sizes (2402.01139v2)

Published 2 Feb 2024 in stat.ML, cs.LG, and stat.ME

Abstract: We introduce a method for online conformal prediction with decaying step sizes. Like previous methods, ours possesses a retrospective guarantee of coverage for arbitrary sequences. However, unlike previous methods, we can simultaneously estimate a population quantile when it exists. Our theory and experiments indicate substantially improved practical properties: in particular, when the distribution is stable, the coverage is close to the desired level for every time point, not just on average over the observed sequence.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (33)
  1. Conformal prediction: A gentle introduction. Foundations and Trends® in Machine Learning, 16(4):494–591, 2023.
  2. Conformal PID control for time series prediction. In Neural Information Processing Systems, 2023.
  3. Conformal prediction beyond exchangeability. arXiv:2202.13415, 2022.
  4. Practical adversarial multivalid conformal prediction. Advances in Neural Information Processing Systems, 35:29362–29373, 2022.
  5. Improved online conformal prediction via strongly adaptive online learning. arXiv preprint arXiv:2302.07869, 2023.
  6. Sequential quantile prediction of time series. IEEE Transactions on Information Theory, 57(3):1664–1674, 2011.
  7. Prediction, learning, and games. Cambridge university press, 2006.
  8. Exact and robust conformal inference methods for predictive machine learning with dependent data. In Conference On Learning Theory, pages 732–749. PMLR, 2018.
  9. Evaluation of individual and ensemble probabilistic forecasts of covid-19 mortality in the united states. Proceedings of the National Academy of Sciences, 119(15):e2113561119, 2022.
  10. Strongly adaptive online learning. In International Conference on Machine Learning, pages 1405–1411. PMLR, 2015.
  11. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255, 2009.
  12. Rick Durrett. Probability: theory and examples, volume 49. Cambridge university press, 2019.
  13. Improving conditional coverage via orthogonal quantile regression. In Advances in Neural Information Processing Systems, 2021.
  14. Achieving risk control in online learning settings. Transactions on Machine Learning Research, 2023.
  15. Fast and scalable gaussian process modeling with applications to astronomical time series. The Astronomical Journal, 154(6):220, 2017.
  16. Asymptotic calibration. Biometrika, 85(2):379–390, 1998.
  17. Adaptive conformal inference under distribution shift. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 1660–1672. Curran Associates, Inc., 2021.
  18. Conformal inference for online prediction with arbitrary distribution shifts. arXiv preprint arXiv:2208.08401, 2022.
  19. Adaptive regret for control of time-varying dynamics. In Learning for Dynamics and Control Conference, pages 560–572. PMLR, 2023.
  20. Splice-2 comparative evaluation: Electricity pricing. 1999.
  21. Regression quantiles. Econometrica: Journal of the Econometric Society, 46(1):33–50, 1978.
  22. Conformal prediction with temporal quantile adjustments. Advances in Neural Information Processing Systems, 35:31017–31030, 2022.
  23. Safe planning in dynamic environments using conformal prediction. IEEE Robotics and Automation Letters, 2023.
  24. Per Aslak Mykland. Financial options and statistical prediction intervals. The Annals of Statistics, 31(5):1413–1438, 2003.
  25. High-dimensional unbiased prediction for sequential decision making. In OPT 2023: Optimization for Machine Learning, 2023.
  26. JA Robinson. Sequential choice of an optimal dose: A prediction intervals approach. Biometrika, 65(1):75–78, 1978.
  27. Methods to compute prediction intervals: A review and new results. Statistical Science, 37(4):580–597, 2022.
  28. Vladimir Vovk. On-line confidence machines are well-calibrated. In The 43rd Annual IEEE Symposium on Foundations of Computer Science, pages 187–196. IEEE, 2002.
  29. Algorithmic Learning in a Random World. Springer, 2005. doi: 10.1007/b106715.
  30. Conformal prediction interval for dynamic time-series. In International Conference on Machine Learning, pages 11559–11569. PMLR, 2021.
  31. Sequential predictive conformal inference for time series. In International Conference on Machine Learning, pages 38707–38727. PMLR, 2023.
  32. Adaptive conformal predictions for time series. In International Conference on Machine Learning, pages 25834–25866. PMLR, 2022.
  33. Martin Zinkevich. Online convex programming and generalized infinitesimal gradient ascent. In Proceedings of the 20th international conference on machine learning (icml-03), pages 928–936, 2003.
Citations (11)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper introduces a novel online conformal prediction method with decaying step sizes that refines prediction thresholds at each time point.
  • Methodology offers dual analytical guarantees, ensuring stability under both adversarial and I.I.D. data scenarios.
  • Empirical tests on Elec2 and Imagenet datasets confirm superior performance over fixed step size methods in maintaining target coverage.

Introduction to the Methodology

A novel method for online conformal prediction with decaying step sizes has been introduced, presenting a significant advancement in the field of online uncertainty quantification for time-series forecasting. This approach stands apart by simultaneously offering retrospective coverage guarantees for arbitrary data sequences and facilitating the estimation of population quantiles under stable distributions. The proposed decaying step sizes contribute to more finely adjusted coverage levels for each time point, moving beyond the average coverage across an observed sequence.

Theoretical Guarantees and Comparisons

Analytical results assert dual guarantees for this method: a worst-case guarantee ensures the convergence of historical miscoverage rates to a predefined error rate even with non-identically distributed data points, while a best-case guarantee entails optimal prediction set convergence for independent and identically distributed (I.I.D.) sequences. These guarantees are not mutual in preceding models, emphasizing the unique positioning of the introduced algorithm within the existing literature spectrum. The ability to adjust to both adversarial and I.I.D. scenarios is grounded in the decaying step sizes, which are tailored through a rigorous mathematical formulation.

Algorithmic Innovations and Conformal Prediction

The methodological innovations lie both in the nuanced updates to the prediction threshold and within the underlying quantile loss function. Unlike fixed step sizes previously examined in online conformal prediction, the unique mathematical framework allows for variable step sizes, adding resilience to distributional shifts and facilitating swift recovery of coverage accuracy post any shifts. At its core, the algorithm adopts principles from conformal prediction, a domain that has garnered attention over the last century due to its broad applicability across various fields such as medicine, robotics, finance, and epidemiology.

Empirical Validation and Performance

The methodology has been verified through empirical assessments on the Elec2 dataset and Imagenet data, illuminating the practical efficacy of the proposed method. Specifically, the decaying step size procedure demonstrated superior stabilization of coverage and threshold values compared to fixed step sizes. Despite the inherently oscillatory behavior of constant step size methods, the decaying step size managed to retain coverage close to the desired confidence level, confirming its practical superiority when faced with both exchangeable data and sequences manifesting considerable distributional changes.

Final Thoughts and Future Directions

The paper presents a promising step toward bridging the gap between online conformal prediction and online learning. The use of decaying step sizes, a common practice in online gradient descent and online learning, reflects a thoughtful integration of robust optimization techniques into the field of online uncertainty quantification. Looking ahead, extending this work to tackle online risk control, accommodate non-I.I.D. time-series, and explore the nexus between online learning strategies and online conformal prediction presents a compelling trajectory for future research.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com