Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Knowledge-tuning Large Language Models with Structured Medical Knowledge Bases for Reliable Response Generation in Chinese (2309.04175v1)

Published 8 Sep 2023 in cs.CL and cs.AI

Abstract: LLMs have demonstrated remarkable success in diverse NLP tasks in general domains. However, LLMs sometimes generate responses with the hallucination about medical facts due to limited domain knowledge. Such shortcomings pose potential risks in the utilization of LLMs within medical contexts. To address this challenge, we propose knowledge-tuning, which leverages structured medical knowledge bases for the LLMs to grasp domain knowledge efficiently and facilitate reliable response generation. We also release cMedKnowQA, a Chinese medical knowledge question-answering dataset constructed from medical knowledge bases to assess the medical knowledge proficiency of LLMs. Experimental results show that the LLMs which are knowledge-tuned with cMedKnowQA, can exhibit higher levels of accuracy in response generation compared with vanilla instruction-tuning and offer a new reliable way for the domain adaptation of LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (12)
  1. Haochun Wang (17 papers)
  2. Sendong Zhao (31 papers)
  3. Zewen Qiang (7 papers)
  4. Zijian Li (71 papers)
  5. Nuwa Xi (11 papers)
  6. Yanrui Du (11 papers)
  7. Haoqiang Guo (3 papers)
  8. Yuhan Chen (39 papers)
  9. Haoming Xu (6 papers)
  10. Bing Qin (186 papers)
  11. Ting Liu (329 papers)
  12. Muzhen Cai (5 papers)
Citations (13)