Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

An Experimental Evaluation of Transformer-based Language Models in the Biomedical Domain (2012.15419v1)

Published 31 Dec 2020 in cs.CL and cs.LG

Abstract: With the growing amount of text in health data, there have been rapid advances in large pre-trained models that can be applied to a wide variety of biomedical tasks with minimal task-specific modifications. Emphasizing the cost of these models, which renders technical replication challenging, this paper summarizes experiments conducted in replicating BioBERT and further pre-training and careful fine-tuning in the biomedical domain. We also investigate the effectiveness of domain-specific and domain-agnostic pre-trained models across downstream biomedical NLP tasks. Our finding confirms that pre-trained models can be impactful in some downstream NLP tasks (QA and NER) in the biomedical domain; however, this improvement may not justify the high cost of domain-specific pre-training.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Paul Grouchy (3 papers)
  2. Shobhit Jain (20 papers)
  3. Michael Liu (23 papers)
  4. Kuhan Wang (2 papers)
  5. Max Tian (2 papers)
  6. Nidhi Arora (1 paper)
  7. Hillary Ngai (4 papers)
  8. Faiza Khan Khattak (10 papers)
  9. Elham Dolatabadi (20 papers)
  10. Sedef Akinli Kocak (5 papers)
Citations (4)