Papers
Topics
Authors
Recent
2000 character limit reached

Fictitious Synthetic Data Can Improve LLM Factuality via Prerequisite Learning (2410.19290v1)

Published 25 Oct 2024 in cs.CL

Abstract: Recent studies have identified one aggravating factor of LLM hallucinations as the knowledge inconsistency between pre-training and fine-tuning, where unfamiliar fine-tuning data mislead the LLM to fabricate plausible but wrong outputs. In this paper, we propose a novel fine-tuning strategy called Prereq-Tune to address this knowledge inconsistency and reduce hallucinations. Fundamentally, Prereq-Tune disentangles the learning of skills and knowledge, so the model learns only the task skills without being impacted by the knowledge inconsistency. To achieve this, Prereq-Tune introduces an additional prerequisite learning stage to learn the necessary knowledge for SFT, allowing subsequent SFT to focus only on task skills. Prereq-Tune can also be combined with fictitious synthetic data to enhance the grounding of LLM outputs to their internal knowledge. Experiments show that Prereq-Tune outperforms existing baselines in improving LLM's factuality across short QA and long-form generation tasks. It also opens new possibilities for knowledge-controlled generation in LLMs. Our code is available at https://github.com/UCSB-NLP-Chang/Prereq_tune.git.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (53)
  1. Self-RAG: Learning to retrieve, generate, and critique through self-reflection. In The Twelfth International Conference on Learning Representations, 2024.
  2. The internal state of an LLM knows when it’s lying. In Houda Bouamor, Juan Pino, and Kalika Bali (eds.), Findings of the Association for Computational Linguistics: EMNLP 2023, December 2023.
  3. INSIDE: LLMs’ internal states retain the power of hallucination detection. In The Twelfth International Conference on Learning Representations, 2024.
  4. Can ai assistants know what they don’t know?, 2024.
  5. Factool: Factuality detection in generative ai – a tool augmented framework for multi-task and multi-domain scenarios, 2023.
  6. Dola: Decoding by contrasting layers improves factuality in large language models. In The Twelfth International Conference on Learning Representations, 2024.
  7. Chain-of-verification reduces hallucination in large language models. In Lun-Wei Ku, Andre Martins, and Vivek Srikumar (eds.), Findings of the Association for Computational Linguistics ACL 2024, August 2024.
  8. Knowledge card: Filling LLMs’ knowledge gaps with plug-in specialized language models. In The Twelfth International Conference on Learning Representations, 2024.
  9. Leo Gao. Behavior cloning is miscalibrated, 2021.
  10. RARR: Researching and revising what language models say, using language models. In Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki (eds.), Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), July 2023.
  11. Does fine-tuning llms on new knowledge encourage hallucinations?, 2024.
  12. Understanding finetuning for factual knowledge extraction, 2024.
  13. Yoav Goldberg. Reinforcement learning for language models, 2023.
  14. The false promise of imitating proprietary language models. In The Twelfth International Conference on Learning Representations, 2024.
  15. TarGEN: Targeted data generation with large language models. In First Conference on Language Modeling, 2024.
  16. Language models can teach themselves to program better. In The Eleventh International Conference on Learning Representations, 2023.
  17. A probabilistic framework for llm hallucination detection via belief tree propagation, 2024.
  18. LoRA: Low-rank adaptation of large language models. In International Conference on Learning Representations, 2022.
  19. A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions, 2023.
  20. Survey of hallucination in natural language generation. ACM Comput. Surv., 2023.
  21. Teaching language models to hallucinate less with synthetic tasks. In The Twelfth International Conference on Learning Representations, 2024.
  22. Unfamiliar finetuning examples control how language models hallucinate, 2024.
  23. Instruct-skillmix: A powerful pipeline for llm instruction tuning, 2024.
  24. Factuality enhanced language models for open-ended text generation. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho (eds.), Advances in Neural Information Processing Systems, 2022.
  25. Chain of natural language inference for reducing large language model ungrounded hallucinations, 2023.
  26. Synthetic data (almost) from scratch: Generalized instruction tuning for language models, 2024.
  27. HaluEval: A large-scale hallucination evaluation benchmark for large language models. In Houda Bouamor, Juan Pino, and Kalika Bali (eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, December 2023a.
  28. Inference-time intervention: Eliciting truthful answers from a language model. In Thirty-seventh Conference on Neural Information Processing Systems, 2023b.
  29. Flame: Factuality-aware alignment for large language models, 2024.
  30. Meta Llama Team. The llama 3 herd of models, 2024.
  31. When not to trust language models: Investigating effectiveness of parametric and non-parametric memories. In Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki (eds.), Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), July 2023.
  32. SelfCheckGPT: Zero-resource black-box hallucination detection for generative large language models. In Houda Bouamor, Juan Pino, and Kalika Bali (eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, December 2023.
  33. Teaching language models to support answers with verified quotes, 2022.
  34. FActScore: Fine-grained atomic evaluation of factual precision in long form text generation. In Houda Bouamor, Juan Pino, and Kalika Bali (eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, December 2023.
  35. Fine-grained hallucination detection and editing for language models. In First Conference on Language Modeling, 2024.
  36. Webgpt: Browser-assisted question-answering with human feedback, 2022.
  37. Direct preference optimization: Your language model is secretly a reward model. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  38. John Schulman. Reinforcement learning from human feedback: Progress and challenges, 2023.
  39. Proximal policy optimization algorithms, 2017.
  40. Retrieval augmentation reduces hallucination in conversation. In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih (eds.), Findings of the Association for Computational Linguistics: EMNLP 2021, November 2021.
  41. Lab: Large-scale alignment for chatbots, 2024.
  42. Principle-driven self-alignment of language models from scratch with minimal human supervision. In Thirty-seventh Conference on Neural Information Processing Systems, 2023a.
  43. Recitation-augmented language models. In The Eleventh International Conference on Learning Representations, 2023b.
  44. SALMON: Self-alignment with instructable reward models. In The Twelfth International Conference on Learning Representations, 2024.
  45. Fine-tuning language models for factuality. In The Twelfth International Conference on Learning Representations, 2024.
  46. Self-instruct: Aligning language models with self-generated instructions. In Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki (eds.), Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, Canada, 2023.
  47. Neural text generation with unlikelihood training. In International Conference on Learning Representations, 2020.
  48. Lilian Weng. Extrinsic hallucinations in llms. lilianweng.github.io, Jul 2024.
  49. Mitigating llm hallucinations via conformal abstention, 2024.
  50. Alignment for honesty, 2023.
  51. Metamath: Bootstrap your own mathematical questions for large language models. In The Twelfth International Conference on Learning Representations, 2024.
  52. R-tuning: Instructing large language models to say ‘I don’t know’. In Kevin Duh, Helena Gomez, and Steven Bethard (eds.), Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), June 2024.
  53. LIMA: Less is more for alignment. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.