Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Quantifying and Attributing the Hallucination of Large Language Models via Association Analysis (2309.05217v1)

Published 11 Sep 2023 in cs.AI and cs.CL

Abstract: Although demonstrating superb performance on various NLP tasks, LLMs still suffer from the hallucination problem, which threatens the reliability of LLMs. To measure the level of hallucination of LLMs, previous works first categorize the hallucination according to the phenomenon similarity, then quantify the proportion that model outputs contain hallucinatory contents. However, such hallucination rates could easily be distorted by confounders. Moreover, such hallucination rates could not reflect the reasons for the hallucination, as similar hallucinatory phenomena may originate from different sources. To address these issues, we propose to combine the hallucination level quantification and hallucination reason investigation through an association analysis, which builds the relationship between the hallucination rate of LLMs with a set of risk factors. In this way, we are able to observe the hallucination level under each value of each risk factor, examining the contribution and statistical significance of each risk factor, meanwhile excluding the confounding effect of other factors. Additionally, by recognizing the risk factors according to a taxonomy of model capability, we reveal a set of potential deficiencies in commonsense memorization, relational reasoning, and instruction following, which may further provide guidance for the pretraining and supervised fine-tuning process of LLMs to mitigate the hallucination.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Li Du (72 papers)
  2. Yequan Wang (44 papers)
  3. Xingrun Xing (13 papers)
  4. Yiqun Ya (1 paper)
  5. Xiang Li (1002 papers)
  6. Xin Jiang (242 papers)
  7. Xuezhi Fang (11 papers)
Citations (10)