Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Reconfidencing LLMs from the Grouping Loss Perspective (2402.04957v3)

Published 7 Feb 2024 in cs.CL

Abstract: LLMs, including ChatGPT and LLaMA, are susceptible to generating hallucinated answers in a confident tone. While efforts to elicit and calibrate confidence scores have proven useful, recent findings show that controlling uncertainty must go beyond calibration: predicted scores may deviate significantly from the actual posterior probabilities due to the impact of grouping loss. In this work, we construct a new evaluation dataset derived from a knowledge base to assess confidence scores given to answers of Mistral and LLaMA. Experiments show that they tend to be overconfident. Further, we show that they are more overconfident on some answers than others, \emph{eg} depending on the nationality of the person in the query. In uncertainty-quantification theory, this is grouping loss. To address this, we propose a solution to reconfidence LLMs, canceling not only calibration but also grouping loss. The LLMs, after the reconfidencing process, indicate improved confidence alignment with the accuracy of their responses.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Lihu Chen (12 papers)
  2. Alexandre Perez-Lebel (6 papers)
  3. Fabian M. Suchanek (12 papers)
  4. Gaƫl Varoquaux (87 papers)
Citations (6)
X Twitter Logo Streamline Icon: https://streamlinehq.com