Med42 -- Evaluating Fine-Tuning Strategies for Medical LLMs: Full-Parameter vs. Parameter-Efficient Approaches (2404.14779v1)
Abstract: This study presents a comprehensive analysis and comparison of two predominant fine-tuning methodologies - full-parameter fine-tuning and parameter-efficient tuning - within the context of medical LLMs. We developed and refined a series of LLMs, based on the Llama-2 architecture, specifically designed to enhance medical knowledge retrieval, reasoning, and question-answering capabilities. Our experiments systematically evaluate the effectiveness of these tuning strategies across various well-known medical benchmarks. Notably, our medical LLM Med42 showed an accuracy level of 72% on the US Medical Licensing Examination (USMLE) datasets, setting a new standard in performance for openly available medical LLMs. Through this comparative analysis, we aim to identify the most effective and efficient method for fine-tuning LLMs in the medical domain, thereby contributing significantly to the advancement of AI-driven healthcare applications.
- Clément Christophe (9 papers)
- Praveen K Kanithi (7 papers)
- Prateek Munjal (6 papers)
- Tathagata Raha (13 papers)
- Nasir Hayat (9 papers)
- Ronnie Rajan (7 papers)
- Ahmed Al-Mahrooqi (1 paper)
- Avani Gupta (5 papers)
- Muhammad Umar Salman (4 papers)
- Gurpreet Gosal (9 papers)
- Bhargav Kanakiya (1 paper)
- Charles Chen (8 papers)
- Natalia Vassilieva (11 papers)
- Boulbaba Ben Amor (14 papers)
- Marco AF Pimentel (7 papers)
- Shadab Khan (11 papers)