Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

InternalInspector $I^2$: Robust Confidence Estimation in LLMs through Internal States (2406.12053v1)

Published 17 Jun 2024 in cs.CL

Abstract: Despite their vast capabilities, LLMs often struggle with generating reliable outputs, frequently producing high-confidence inaccuracies known as hallucinations. Addressing this challenge, our research introduces InternalInspector, a novel framework designed to enhance confidence estimation in LLMs by leveraging contrastive learning on internal states including attention states, feed-forward states, and activation states of all layers. Unlike existing methods that primarily focus on the final activation state, InternalInspector conducts a comprehensive analysis across all internal states of every layer to accurately identify both correct and incorrect prediction processes. By benchmarking InternalInspector against existing confidence estimation methods across various natural language understanding and generation tasks, including factual question answering, commonsense reasoning, and reading comprehension, InternalInspector achieves significantly higher accuracy in aligning the estimated confidence scores with the correctness of the LLM's predictions and lower calibration error. Furthermore, InternalInspector excels at HaluEval, a hallucination detection benchmark, outperforming other internal-based confidence estimation methods in this task.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Mohammad Beigi (4 papers)
  2. Ying Shen (76 papers)
  3. Runing Yang (2 papers)
  4. Zihao Lin (22 papers)
  5. Qifan Wang (129 papers)
  6. Ankith Mohan (5 papers)
  7. Jianfeng He (32 papers)
  8. Ming Jin (130 papers)
  9. Chang-Tien Lu (54 papers)
  10. Lifu Huang (91 papers)
Citations (1)