Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Enhancing Uncertainty-Based Hallucination Detection with Stronger Focus (2311.13230v1)

Published 22 Nov 2023 in cs.CL and cs.AI

Abstract: LLMs have gained significant popularity for their impressive performance across diverse fields. However, LLMs are prone to hallucinate untruthful or nonsensical outputs that fail to meet user expectations in many real-world applications. Existing works for detecting hallucinations in LLMs either rely on external knowledge for reference retrieval or require sampling multiple responses from the LLM for consistency verification, making these methods costly and inefficient. In this paper, we propose a novel reference-free, uncertainty-based method for detecting hallucinations in LLMs. Our approach imitates human focus in factuality checking from three aspects: 1) focus on the most informative and important keywords in the given text; 2) focus on the unreliable tokens in historical context which may lead to a cascade of hallucinations; and 3) focus on the token properties such as token type and token frequency. Experimental results on relevant datasets demonstrate the effectiveness of our proposed method, which achieves state-of-the-art performance across all the evaluation metrics and eliminates the need for additional information.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Tianhang Zhang (16 papers)
  2. Lin Qiu (47 papers)
  3. Qipeng Guo (72 papers)
  4. Cheng Deng (67 papers)
  5. Yue Zhang (618 papers)
  6. Zheng Zhang (486 papers)
  7. Chenghu Zhou (55 papers)
  8. Xinbing Wang (98 papers)
  9. Luoyi Fu (41 papers)
Citations (27)