An Experimental Evaluation of Transformer-based Language Models in the Biomedical Domain (2012.15419v1)
Abstract: With the growing amount of text in health data, there have been rapid advances in large pre-trained models that can be applied to a wide variety of biomedical tasks with minimal task-specific modifications. Emphasizing the cost of these models, which renders technical replication challenging, this paper summarizes experiments conducted in replicating BioBERT and further pre-training and careful fine-tuning in the biomedical domain. We also investigate the effectiveness of domain-specific and domain-agnostic pre-trained models across downstream biomedical NLP tasks. Our finding confirms that pre-trained models can be impactful in some downstream NLP tasks (QA and NER) in the biomedical domain; however, this improvement may not justify the high cost of domain-specific pre-training.
- Paul Grouchy (3 papers)
- Shobhit Jain (20 papers)
- Michael Liu (23 papers)
- Kuhan Wang (2 papers)
- Max Tian (2 papers)
- Nidhi Arora (1 paper)
- Hillary Ngai (4 papers)
- Faiza Khan Khattak (10 papers)
- Elham Dolatabadi (20 papers)
- Sedef Akinli Kocak (5 papers)