Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 99 tok/s
Gemini 2.5 Pro 55 tok/s Pro
GPT-5 Medium 23 tok/s
GPT-5 High 19 tok/s Pro
GPT-4o 108 tok/s
GPT OSS 120B 465 tok/s Pro
Kimi K2 179 tok/s Pro
2000 character limit reached

Collaborative decoding of critical tokens for boosting factuality of large language models (2402.17982v1)

Published 28 Feb 2024 in cs.CL

Abstract: The most common training pipeline for LLMs includes pretraining, finetuning and aligning phases, with their respective resulting models, such as the pretrained model and the finetuned model. Finetuned and aligned models show improved abilities of instruction following and safe generation, however their abilities to stay factual about the world are impacted by the finetuning process. Furthermore, the common practice of using sampling during generation also increases chances of hallucination. In this work, we introduce a collaborative decoding framework to harness the high factuality within pretrained models through the concept of critical tokens. We first design a critical token classifier to decide which model to use for the next token, and subsequently generates the next token using different decoding strategies. Experiments with different models and datasets show that our decoding framework is able to reduce model hallucination significantly, showcasing the importance of the collaborative decoding framework.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. Training a helpful and harmless assistant with reinforcement learning from human feedback. ArXiv preprint, abs/2204.05862.
  2. Leo Breiman. 1996. Bagging predictors. Machine learning, 24:123–140.
  3. A survey on evaluation of large language models. ACM Transactions on Intelligent Systems and Technology.
  4. Dola: Decoding by contrasting layers improves factuality in large language models.
  5. Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416.
  6. COLD: A benchmark for Chinese offensive language detection. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 11580–11599, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  7. Chain-of-verification reduces hallucination in large language models. arXiv preprint arXiv:2309.11495.
  8. Retrieval-augmented generation for large language models: A survey.
  9. A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions. arXiv preprint arXiv:2311.05232.
  10. Few-shot learning with retrieval augmented language models. arXiv preprint arXiv:2208.03299.
  11. Mistral 7b. arXiv preprint arXiv:2310.06825.
  12. TriviaQA: A large scale distantly supervised challenge dataset for reading comprehension. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1601–1611. Association for Computational Linguistics.
  13. Natural questions: a benchmark for question answering research. Transactions of the Association of Computational Linguistics.
  14. Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in Neural Information Processing Systems, 33:9459–9474.
  15. The dawn after the dark: An empirical study on factuality hallucination in large language models.
  16. Halueval: A large-scale hallucination evaluation benchmark for large language models. In The 2023 Conference on Empirical Methods in Natural Language Processing.
  17. Inference-time intervention: Eliciting truthful answers from a language model. arXiv preprint arXiv:2306.03341.
  18. Contrastive decoding: Open-ended text generation as optimization. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers).
  19. Truthfulqa: Measuring how models mimic human falsehoods. arXiv preprint arXiv:2109.07958.
  20. Mitigating the alignment tax of rlhf.
  21. Query rewriting for retrieval-augmented large language models. arXiv preprint arXiv:2305.14283.
  22. Factscore: Fine-grained atomic evaluation of factual precision in long form text generation. arXiv preprint arXiv:2305.14251.
  23. Sean O’Brien and Mike Lewis. 2023. Contrastive decoding improves reasoning in large language models. ArXiv preprint, abs/2309.09117.
  24. OpenAI. 2022. Introducing chatgpt. https://openai.com/blog/chatgpt.
  25. OpenAI. 2023. GPT-4 technical report. ArXiv preprint, abs/2303.08774.
  26. David Opitz and Richard Maclin. 1999. Popular ensemble methods: An empirical study. Journal of artificial intelligence research, 11:169–198.
  27. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
  28. The troubling emergence of hallucination in large language models - an extensive definition, quantification, and prescriptive remediations. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 2541–2573, Singapore. Association for Computational Linguistics.
  29. A thorough examination of decoding methods in the era of llms. arXiv preprint arXiv:2402.06925.
  30. Retrieval augmentation reduces hallucination in conversation. arXiv preprint arXiv:2104.07567.
  31. Fine-tuning language models for factuality. arXiv preprint arXiv:2311.08401.
  32. Llama 2: Open foundation and fine-tuned chat models.
  33. Survey on factuality in large language models: Knowledge, retrieval and domain-specificity. arXiv preprint arXiv:2310.07521.
  34. Hallucination is inevitable: An innate limitation of large language models. arXiv preprint arXiv:2401.11817.
  35. Llm lies: Hallucinations are not bugs, but features as adversarial examples. arXiv preprint arXiv:2310.01469.
  36. Chain-of-note: Enhancing robustness in retrieval-augmented language models. arXiv preprint arXiv:2311.09210.
  37. SafeConv: Explaining and correcting conversational unsafe behavior. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 22–35, Toronto, Canada. Association for Computational Linguistics.
  38. Alleviating hallucinations of large language models through induced hallucinations. arXiv preprint arXiv:2312.15710.
  39. Verify-and-edit: A knowledge-enhanced chain-of-thought framework. arXiv preprint arXiv:2305.03268.
  40. Texygen: A benchmarking platform for text generation models. In The 41st international ACM SIGIR conference on research & development in information retrieval, pages 1097–1100.
Citations (4)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube