Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

KnowTuning: Knowledge-aware Fine-tuning for Large Language Models (2402.11176v3)

Published 17 Feb 2024 in cs.CL and cs.AI

Abstract: Despite their success at many NLP tasks, LLMs still struggle to effectively leverage knowledge for knowledge-intensive tasks, manifesting limitations such as generating incomplete, non-factual, or illogical answers. These limitations stem from inadequate knowledge awareness of LLMs during vanilla fine-tuning. To address these problems, we propose a knowledge-aware fine-tuning (KnowTuning) method to improve fine-grained and coarse-grained knowledge awareness of LLMs. We devise a fine-grained knowledge augmentation stage to train LLMs to identify difficult fine-grained knowledge in answers. We also propose a coarse-grained knowledge comparison stage to train LLMs to distinguish between reliable and unreliable knowledge, in three aspects: completeness, factuality, and logicality. Extensive experiments on both generic and medical question answering (QA) datasets confirm the effectiveness of KnowTuning, through automatic and human evaluations, across various sizes of LLMs. We further verify that KnowTuning generates more facts with less factual error rate under fine-grained facts evaluation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Yougang Lyu (11 papers)
  2. Lingyong Yan (29 papers)
  3. Shuaiqiang Wang (68 papers)
  4. Haibo Shi (9 papers)
  5. Dawei Yin (165 papers)
  6. Pengjie Ren (95 papers)
  7. Zhumin Chen (78 papers)
  8. Maarten de Rijke (261 papers)
  9. Zhaochun Ren (117 papers)
Citations (4)