Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MKRAG: Medical Knowledge Retrieval Augmented Generation for Medical Question Answering (2309.16035v3)

Published 27 Sep 2023 in cs.CL and cs.AI

Abstract: LLMs, although powerful in general domains, often perform poorly on domain-specific tasks such as medical question answering (QA). In addition, LLMs tend to function as "black-boxes", making it challenging to modify their behavior. To address the problem, our work employs a transparent process of retrieval augmented generation (RAG), aiming to improve LLM responses without the need for fine-tuning or retraining. Specifically, we propose a comprehensive retrieval strategy to extract medical facts from an external knowledge base, and then inject them into the LLM's query prompt. Focusing on medical QA, we evaluate the impact of different retrieval models and the number of facts on LLM performance using the MedQA-SMILE dataset. Notably, our retrieval-augmented Vicuna-7B model exhibited an accuracy improvement from 44.46% to 48.54%. This work underscores the potential of RAG to enhance LLM performance, offering a practical approach to mitigate the challenges posed by black-box LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Yucheng Shi (30 papers)
  2. Shaochen Xu (16 papers)
  3. Zhengliang Liu (91 papers)
  4. Tianming Liu (161 papers)
  5. Xiang Li (1002 papers)
  6. Ninghao Liu (98 papers)
  7. Tianze Yang (13 papers)
  8. Quanzheng Li (122 papers)
Citations (12)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets