Stronger Calibration Lower Bounds via Sidestepping
Abstract: We consider an online binary prediction setting where a forecaster observes a sequence of $T$ bits one by one. Before each bit is revealed, the forecaster predicts the probability that the bit is $1$. The forecaster is called well-calibrated if for each $p \in [0, 1]$, among the $n_p$ bits for which the forecaster predicts probability $p$, the actual number of ones, $m_p$, is indeed equal to $p \cdot n_p$. The calibration error, defined as $\sum_p |m_p - p n_p|$, quantifies the extent to which the forecaster deviates from being well-calibrated. It has long been known that an $O(T{2/3})$ calibration error is achievable even when the bits are chosen adversarially, and possibly based on the previous predictions. However, little is known on the lower bound side, except an $\Omega(\sqrt{T})$ bound that follows from the trivial example of independent fair coin flips. In this paper, we prove an $\Omega(T{0.528})$ bound on the calibration error, which is the first super-$\sqrt{T}$ lower bound for this setting to the best of our knowledge. The technical contributions of our work include two lower bound techniques, early stopping and sidestepping, which circumvent the obstacles that have previously hindered strong calibration lower bounds. We also propose an abstraction of the prediction setting, termed the Sign-Preservation game, which may be of independent interest. This game has a much smaller state space than the full prediction setting and allows simpler analyses. The $\Omega(T{0.528})$ lower bound follows from a general reduction theorem that translates lower bounds on the game value of Sign-Preservation into lower bounds on the calibration error.
- Glenn W Brier. Verification of forecasts expressed in terms of probability. Monthly Weather Review, 78(1):1–3, 1950.
- A Philip Dawid. The well-calibrated bayesian. Journal of the American Statistical Association, 77(379):605–610, 1982.
- Forecast-hedging and calibration. http://www.ma.huji.ac.il/hart/papers/calib-int.pdf, 2020. Accessed: 2020-12-01.
- An easier way to calibrate. Games and Economic Behavior, 29(1-2):131–137, 1999.
- Dean P Foster. A proof of calibration via blackwell’s approachability theorem. Games and Economic Behavior, 29(1-2):73–78, 1999.
- Complexity-based approach to calibration with checking rules. In Conference on Learning Theory (COLT), pages 293–314, 2011.
- Asymptotic calibration. Biometrika, 85(2):379–390, 1998.
- On calibration of modern neural networks. In International Conference on Machine Learning (ICML), pages 1321–1330, 2017.
- Sergiu Hart. Calibrated forecasts: The minimax proof. http://www.ma.huji.ac.il/hart/papers/calib-minmax.pdf, 2020. Accessed: 2020-12-01.
- Calibration for the (computationally-identifiable) masses. In International Conference on Machine Learning (ICML), pages 1939–1948, 2018.
- Moment multicalibration for uncertainty estimation. arXiv preprint arXiv:2008.08037, 2020.
- Calibrated structured prediction. In Advances in Neural Information Processing Systems (NIPS), pages 3474–3482, 2015.
- Verified uncertainty calibration. In Advances in Neural Information Processing Systems (NeurIPS), pages 3792–3803, 2019.
- Inherent trade-offs in the fair determination of risk scores. In Innovations in Theoretical Computer Science Conference (ITCS), pages 43:1–43:23, 2017.
- Ehud Lehrer. Any inspection is manipulable. Econometrica, 69(5):1333–1347, 2001.
- On fairness and calibration. In Advances in Neural Information Processing Systems (NIPS), pages 5680–5689, 2017.
- With malice towards none: Assessing uncertainty via equalized coverage. arXiv preprint arXiv:1908.05428, 2019.
- Online learning: Beyond regret. In Conference on Learning Theory (COLT), pages 559–594, 2011.
- Sample complexity of uniform convergence for multicalibration. arXiv preprint arXiv:2005.01757, 2020.
- Calibration with many checking rules. Mathematics of Operations Research, 28(1):141–153, 2003.
- Vladimir Vovk. Non-asymptotic calibration and resolution. Theoretical Computer Science, 387(1):77–89, 2007.
- Individual calibration with randomized forecasting. In International Conference on Machine Learning (ICML), pages 8366–8376, 2020.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.