Papers
Topics
Authors
Recent
2000 character limit reached

LURE: Mitigating Hallucinations in LVLMs

Updated 15 December 2025
  • LURE is a modular toolkit that mitigates object and factual hallucinations in LVLMs using targeted post-hoc interventions and architectural adjustments.
  • It employs causal transformer analysis and attention head reweighting to identify and suppress erroneous outputs during image captioning.
  • The approach integrates statistical masking and lightweight revisor networks, achieving measurable improvements on multiple LVLM benchmarks.

The LVLM Hallucination Revisor (LURE) denotes a suite of post-hoc and architecture-level interventions for mitigating object and factual hallucinations in Large Vision-LLMs (LVLMs). LURE algorithms generally operate by identifying and suppressing model behaviors that cause the model to describe entities not present in the input image or context. Approaches falling under LURE span statistical postprocessing, attention head reweighting based on causal analysis, and lightweight revisor networks for caption rewriting. These methods are universally designed to be modular and can be retrofitted to arbitrary LVLM architectures without requiring retraining of the core model weights (Zhou et al., 2023, Qian et al., 21 Nov 2025).

1. Problem Formulation and Hallucination Factors

In LVLMs, object hallucination is defined as the generation of references to objects or facts that do not correspond to perceptual evidence in the source modality (typically an image). For a generated sequence s=(z1,,zNs)s=(z_1,\ldots,z_{N_s}), the set of objects Os={os,1,...,os,nh+nr}\mathcal{O}_s=\{o_{s,1},...,o_{s,n_h+n_r}\} divides into real objects (present in the image) and hallucinated objects (absent). LURE methods are motivated by the observation that hallucination probability is not uniformly distributed but is influenced by three key statistical factors (Zhou et al., 2023):

  • Co-occurrence Factor: Objects frequently co-mentioned in captions increase hallucination risk. The pairwise co-occurrence metric

C(oi,oj)=S(oi)S(oj)S(oi)+S(oj)C(o_i, o_j) = \frac{| \mathcal{S}(o_i) \cap \mathcal{S}(o_j) |}{|\mathcal{S}(o_i)| + |\mathcal{S}(o_j)|}

is central in analyzing which false objects are more likely to appear.

  • Uncertainty Factor: Hallucinated tokens exhibit higher negative log-probabilities at decode time. For each object,

UnScore(s,i)=logpM(os,is<i,x)\mathrm{UnScore}(s, i) = -\log p_M(o_{s,i} | s_{<i}, x)

  • Positional Factor: Hallucinations manifest predominantly toward the end of generated captions, quantifiable by the normalized position index

PoScore(s,i)=Index(os,i)/Ns\mathrm{PoScore}(s, i) = \mathrm{Index}(o_{s,i}) / N_s

Collectively, these factors provide a basis for identifying high-risk segments for revision.

2. LURE Algorithmic Frameworks

The LURE strategy synthesizes insights from both causal transformer analysis and statistical factors to yield practical revision algorithms (Zhou et al., 2023, Qian et al., 21 Nov 2025).

2.1 Causal Path Analysis and Attention Head Intervention

LVLM generation can be decomposed into three causal pathways:

  • Image-to-Input-Text (I2T-in): Influence of image features on prompt token representations.
  • Image-to-Output-Text (I2T-out): Direct influence of visual tokens during token generation.
  • Text-to-Text (T2T): Influence of prior textual tokens on subsequent outputs.

Each pathway is instantiated in the multi-head attention structure, where attention heads are scored for their contribution to either grounding or hallucinating. The head scores ST2T(l,n)S_{\mathrm{T2T}}^{(l,n)} and SI2T(l,n)S_{\mathrm{I2T}}^{(l,n)} are computed based on their log-probability increases for correct vs. hallucinated tokens and their attention concentration on true versus absent objects, respectively.

Intervention is implemented by rescaling the outputs of identified heads during inference: λ(l,n)={γif (l,n)Z γ+if (l,n)Z+ 1otherwise\lambda^{(l,n)} = \begin{cases} \gamma^- & \text{if } (l,n) \in Z^- \ \gamma^+ & \text{if } (l,n) \in Z^+ \ 1 & \text{otherwise} \end{cases} with γ=0,γ+=2\gamma^-=0, \gamma^+=2 typical default choices.

2.2 Statistical Mask-and-Revise: Caption Revisor Networks

An alternative LURE instantiation utilizes a mask-and-revise sequence, where an auxiliary network is trained to correct hallucinations post-generation (Zhou et al., 2023). The training protocol consists of:

  1. Synthetic Hallucination Pairing: Ground-truth captions yy are corrupted into hh by object insertion based on co-occurrence, followed by masking of high uncertainty and late-positioned objects.
  2. Revisor Training: The revisor RθR_\theta is trained to map (x,h)y(x, h) \rightarrow y using standard autoregressive loss:

θθαθE(x,h,y)[logpRθ(yx,h)]\theta \gets \theta - \alpha \nabla_\theta \mathbb{E}_{(x,h,y)}[-\log p_{R_\theta}(y | x, h)]

At inference, hallucination-prone spans in the generated caption are replaced with a special token (e.g., “[IDK]”) and the revisor network generates a cleaned caption.

3. Integration and Inference Pipelines

LURE is designed to be modular and compatible with black-box LVLMs.

  • Post-hoc Attention Intervention: Attention weights are modified on-the-fly based on statically scored head indices, without altering base model parameters or requiring access to underlying training data (Qian et al., 21 Nov 2025).
  • Revisor-based Caption Rewriting: The revisor model is invoked only after full-sequence generation, leveraging token-level uncertainty and position signaling to localize corrections (Zhou et al., 2023).
  • Thresholds and Hyperparameters: Masking and reweighting thresholds (notably γ\gamma for uncertainty and η\eta for position) are set by cross-validation and significantly impact revision aggressiveness.

The LURE pipeline for attention head reweighting:

1
2
3
4
5
6
7
8
Input: Prompt P, Image I
1. Detect alignment format (discriminative/generative)
2. Load precomputed head scores {S_T2T, S_I2T}
3. Select relevant heads Z^-, Z^+
4. Set scaling λ^(l,n) for each head
5. Run the LVLM, applying λ^(l,n) in each attention layer
6. Decode output sequence y
Output: Revised response y

4. Empirical Evaluation and Performance

Across multiple benchmarks and backbone models, LURE methodologies show robust hallucination mitigation.

Results on Standard LVLM Benchmarks (LLaVA-7B backbone) (Qian et al., 21 Nov 2025):

Method POPE Acc POPE F1 MCQ-POPE Acc MCQ-POPE F1 CHAIR C_S CHAIR C_I
Vanilla 85.1 83.7 72.8 72.9 52.2 14.6
PAI 86.4 85.0 78.0 78.0 28.8 7.9
AD-HH 85.0 83.6 78.5 78.6 33.2 7.5
LURE 87.2 86.0 80.5 80.5 26.6 7.2

On the MME hallucination subset, LURE yields a total score of 600.0 (±8.7), compared to 540.0 baseline.

Key empirical findings:

  • LURE reduces CHAIRS_S and CHAIRI_I (sentence-/instance-level hallucination) on open-ended captioning.
  • Consistent F1 and accuracy improvements on object existence and multiple-choice detection (POPE, MCQ-POPE).
  • Evaluation is model- and task-agnostic, showing transferability across LVLM backbones (e.g., LLaVA, Qwen-VL).

A plausible implication is that targeted, statically determined architectural interventions can outperform previously proposed retraining or beam rescoring methods, while remaining lightweight and easily deployable.

5. Limitations, Generalization, and Extensions

Identified constraints in current LURE methods include:

  • Synthetic Training Pairs: Caption revisors rely on hallucination simulation (e.g., GPT-3.5 generated corruptions). The use of real, human-annotated hallucination data could lead to further improvements (Zhou et al., 2023).
  • Threshold Sensitivity: The effect of γ\gamma and η\eta can be task-dependent, necessitating careful tuning for new deployment domains.
  • Attribute/Relation Hallucinations: Present instantiations focus on object hallucinations; extension to attribute, relational, or scenario-level errors remains an open direction.
  • Beam Search Integration: While LURE is compatible with standard decoders, deeper integration into beam/n-best rescoring could yield further gains.

LURE can be combined with external detectors (e.g., CLIP or specialist object detectors) as part of broader revision frameworks, potentially unifying mask-and-revise, causal analysis, and visual-linguistic consistency scoring (Deng et al., 23 Feb 2024).

6. Conclusions and Outlook

LURE provides a unified, efficient, and empirically validated toolkit for reducing LVLM hallucinations via post-hoc architectural and statistical revision. By intervening along causally motivated paths—particularly by manipulating identified “hallucinatory” and “grounding” attention heads—and supplementing with revisor architectures, LURE achieves measurable reductions in object hallucinations with minimal cost and maximal flexibility. Extensions toward richer annotation, attribute-aware revision, and deeper integration into LVLM inference pipelines represent promising directions for further study (Zhou et al., 2023, Qian et al., 21 Nov 2025).

Whiteboard

Follow Topic

Get notified by email when new papers are published related to LVLM Hallucination Revisor (LURE).