Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Med42 -- Evaluating Fine-Tuning Strategies for Medical LLMs: Full-Parameter vs. Parameter-Efficient Approaches (2404.14779v1)

Published 23 Apr 2024 in cs.CL

Abstract: This study presents a comprehensive analysis and comparison of two predominant fine-tuning methodologies - full-parameter fine-tuning and parameter-efficient tuning - within the context of medical LLMs. We developed and refined a series of LLMs, based on the Llama-2 architecture, specifically designed to enhance medical knowledge retrieval, reasoning, and question-answering capabilities. Our experiments systematically evaluate the effectiveness of these tuning strategies across various well-known medical benchmarks. Notably, our medical LLM Med42 showed an accuracy level of 72% on the US Medical Licensing Examination (USMLE) datasets, setting a new standard in performance for openly available medical LLMs. Through this comparative analysis, we aim to identify the most effective and efficient method for fine-tuning LLMs in the medical domain, thereby contributing significantly to the advancement of AI-driven healthcare applications.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (16)
  1. Clément Christophe (9 papers)
  2. Praveen K Kanithi (7 papers)
  3. Prateek Munjal (6 papers)
  4. Tathagata Raha (13 papers)
  5. Nasir Hayat (9 papers)
  6. Ronnie Rajan (7 papers)
  7. Ahmed Al-Mahrooqi (1 paper)
  8. Avani Gupta (5 papers)
  9. Muhammad Umar Salman (4 papers)
  10. Gurpreet Gosal (9 papers)
  11. Bhargav Kanakiya (1 paper)
  12. Charles Chen (8 papers)
  13. Natalia Vassilieva (11 papers)
  14. Boulbaba Ben Amor (14 papers)
  15. Marco AF Pimentel (7 papers)
  16. Shadab Khan (11 papers)
Citations (13)
X Twitter Logo Streamline Icon: https://streamlinehq.com