Robustness of decreasing calibration error near perfect accuracy in underconfident settings
Investigate whether the observed decrease in calibration error as semantic accuracy approaches 100% for configurations that are systematically underconfident is a robust phenomenon across models, datasets, and prompting setups.
References
For the underconfident configurations, we see also little correlation overall, except for in the high-accuracy regime: calibration error tends to decrease when models approach perfect semantic accuracy. However, it is not clear whether this is a robust phenomenon.
— Trained on Tokens, Calibrated on Concepts: The Emergence of Semantic Calibration in LLMs
(2511.04869 - Nakkiran et al., 6 Nov 2025) in Section 5: Experiments — Model Scaling Effects