Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Parameter-Efficient Fine-Tuning of LLaMA for the Clinical Domain (2307.03042v3)

Published 6 Jul 2023 in cs.CL and cs.LG

Abstract: Adapting pretrained LLMs to novel domains, such as clinical applications, traditionally involves retraining their entire set of parameters. Parameter-Efficient Fine-Tuning (PEFT) techniques for fine-tuning LLMs significantly reduce computational requirements by selectively fine-tuning small subsets of parameters. In this study, we propose a two-step PEFT framework and evaluate it in the clinical domain. Our approach combines a specialised PEFT adapter layer designed for clinical domain adaptation with another adapter specialised for downstream tasks. We evaluate the framework on multiple clinical outcome prediction datasets, comparing it to clinically trained LLMs. Our framework achieves a better AUROC score averaged across all clinical downstream tasks compared to clinical LLMs. In particular, we observe large improvements of 4-5% AUROC in large-scale multilabel classification tasks, such as diagnoses and procedures classification. To our knowledge, this study is the first to provide an extensive empirical analysis of the interplay between PEFT techniques and domain adaptation in an important real-world domain of clinical applications.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. Publicly available clinical BERT embeddings. In Proceedings of the 2nd Clinical Natural Language Processing Workshop, pages 72–78, Minneapolis, Minnesota, USA. Association for Computational Linguistics.
  2. BERT: pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Volume 1 (Long and Short Papers), pages 4171–4186. Association for Computational Linguistics.
  3. Peter I. Frazier. 2018. A tutorial on bayesian optimization. CoRR, abs/1807.02811.
  4. Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing. ACM Transactions on Computing for Healthcare, 3(1):1–23.
  5. Parameter-efficient transfer learning for nlp. In Proceedings of the 36th International Conference on Machine Learning, page 2790–2799. PMLR.
  6. LoRA: Low-rank adaptation of large language models. In International Conference on Learning Representations.
  7. Clinical XLNet: Modeling Sequential Clinical Notes and Predicting Prolonged Mechanical Ventilation. In Proceedings of the 3rd Clinical Natural Language Processing Workshop, pages 94–100, Online. Association for Computational Linguistics.
  8. MIMIC-IV, a freely accessible electronic health record dataset. Scientific Data, 10(1):1.
  9. MIMIC-III, a freely accessible critical care database. Scientific Data, 3(1):160035.
  10. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics, 36(4):1234–1240.
  11. Eric Lehman and Alistair Johnson. 2023. Clinical-T5: Large Language Models Built Using MIMIC Clinical Text.
  12. The power of scale for parameter-efficient prompt tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 3045–3059, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  13. Xiang Lisa Li and Percy Liang. 2021. Prefix-Tuning: Optimizing Continuous Prompts for Generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4582–4597, Online. Association for Computational Linguistics.
  14. Clinical-longformer and clinical-bigbird: Transformers for long clinical sequences. CoRR, abs/2201.11838.
  15. Beyond one-model-fits-all: A survey of domain specialization for large language models. CoRR, abs/2305.18703.
  16. P-tuning v2: Prompt tuning can be comparable to fine-tuning universally across scales and tasks. CoRR, abs/2110.07602.
  17. GPT understands, too. CoRR, abs/2103.10385.
  18. BioGPT: generative pre-trained transformer for biomedical text generation and mining. Briefings in Bioinformatics, 23(6). Bbac409.
  19. Peft: State-of-the-art parameter-efficient fine-tuning methods. https://github.com/huggingface/peft.
  20. UmlsBERT: Clinical domain knowledge augmentation of contextual embeddings using the Unified Medical Language System Metathesaurus. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 1744–1753, Online. Association for Computational Linguistics.
  21. Transfer learning in biomedical natural language processing: An evaluation of BERT and ELMo on ten benchmarking datasets. In Proceedings of the 18th BioNLP Workshop and Shared Task, pages 58–65, Florence, Italy. Association for Computational Linguistics.
  22. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21(140):1–67. Citation Key: JMLR:v21:20-074.
  23. Modular and Parameter-Efficient Fine-Tuning for NLP Models. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: Tutorial Abstracts, pages 23–29, Abu Dubai, UAE. Association for Computational Linguistics.
  24. The UMLS Metathesaurus: Representing different views of biomedical concepts. Bulletin of the Medical Library Association, 81(2):217–222.
  25. Large language models encode clinical knowledge. CoRR, abs/2212.13138.
  26. Towards expert-level medical question answering with large language models. CoRR, abs/2305.09617.
  27. Llama: Open and efficient foundation language models. CoRR, abs/2302.13971.
  28. Clinical Outcome Prediction from Admission Notes using Self-Supervised Knowledge Integration. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 881–893, Online. Association for Computational Linguistics.
  29. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net.
  30. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online. Association for Computational Linguistics.
  31. Pmc-llama: Further finetuning llama on medical papers. CoRR, abs/2304.14454.
  32. A large language model for electronic health records. npj Digit. Medicine, 5.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Aryo Pradipta Gema (18 papers)
  2. Luke Daines (4 papers)
  3. Pasquale Minervini (88 papers)
  4. Beatrice Alex (21 papers)
  5. Tom Hope (41 papers)
Citations (32)