Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SH2: Self-Highlighted Hesitation Helps You Decode More Truthfully (2401.05930v4)

Published 11 Jan 2024 in cs.CL and cs.AI

Abstract: LLMs demonstrate great performance in text generation. However, LLMs are still suffering from hallucinations. In this work, we propose an inference-time method, Self-Highlighted Hesitation (SH2), to help LLMs decode more truthfully. SH2 is based on a simple fact rooted in information theory that for an LLM, the tokens predicted with lower probabilities are prone to be more informative than others. Our analysis shows that the tokens assigned with lower probabilities by an LLM are more likely to be closely related to factual information, such as nouns, proper nouns, and adjectives. Therefore, we propose to ''highlight'' the factual information by selecting the tokens with the lowest probabilities and concatenating them to the original context, thus forcing the model to repeatedly read and hesitate on these tokens before generation. During decoding, we also adopt contrastive decoding to emphasize the difference in the output probabilities brought by the hesitation. Experimental results demonstrate that our SH2, requiring no additional data or models, can effectively help LLMs elicit factual knowledge and distinguish hallucinated contexts. Significant and consistent improvements are achieved by SH2 for LLaMA-7b, LLaMA2-7b and Mistral-7b on multiple hallucination tasks.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (24)
  1. Dola: Decoding by contrasting layers improves factuality in large language models. CoRR, abs/2309.03883.
  2. Rarr: Researching and revising what language models say, using language models.
  3. Critic: Large language models can self-correct with tool-interactive critiquing.
  4. Think before you speak: Training language models with pause tokens. arXiv preprint arXiv:2310.02226.
  5. Alex Graves. 2017. Adaptive computation time for recurrent neural networks.
  6. Teaching machines to read and comprehend.
  7. Halueval: A large-scale hallucination evaluation benchmark for large language models.
  8. Inference-time intervention: Eliciting truthful answers from a language model.
  9. Contrastive decoding: Open-ended text generation as optimization. In ACL (1), pages 12286–12312. Association for Computational Linguistics.
  10. Truthfulqa: Measuring how models mimic human falsehoods. In ACL (1), pages 3214–3252. Association for Computational Linguistics.
  11. Generating benchmarks for factuality evaluation of language models. CoRR, abs/2307.06908.
  12. Sean O’Brien and Mike Lewis. 2023. Contrastive decoding improves reasoning in large language models. CoRR, abs/2309.09117.
  13. OpenAI. 2023. Gpt-4 technical report.
  14. Check your facts and try again: Improving large language models with external knowledge and automated feedback.
  15. Trusting your evidence: Hallucinate less with context-aware decoding. CoRR, abs/2305.14739.
  16. Autoprompt: Eliciting knowledge from language models with automatically generated prompts.
  17. Moss: Training conversational language models from synthetic data.
  18. Stanford alpaca: An instruction-following llama model. https://github.com/tatsu-lab/stanford_alpaca.
  19. Llama: Open and efficient foundation language models. CoRR, abs/2302.13971.
  20. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288.
  21. Chain-of-thought prompting elicits reasoning in large language models. In NeurIPS.
  22. Measuring association between labels and free-text rationales.
  23. How language model hallucinations can snowball. CoRR, abs/2305.13534.
  24. Siren’s song in the ai ocean: A survey on hallucination in large language models. arXiv preprint arXiv:2309.01219.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Jushi Kai (5 papers)
  2. Hai Hu (23 papers)
  3. Zhouhan Lin (57 papers)
  4. Tianhang Zhang (16 papers)
Citations (7)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets