Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

BioBERT: a pre-trained biomedical language representation model for biomedical text mining (1901.08746v4)

Published 25 Jan 2019 in cs.CL
BioBERT: a pre-trained biomedical language representation model for biomedical text mining

Abstract: Biomedical text mining is becoming increasingly important as the number of biomedical documents rapidly grows. With the progress in NLP, extracting valuable information from biomedical literature has gained popularity among researchers, and deep learning has boosted the development of effective biomedical text mining models. However, directly applying the advancements in NLP to biomedical text mining often yields unsatisfactory results due to a word distribution shift from general domain corpora to biomedical corpora. In this article, we investigate how the recently introduced pre-trained LLM BERT can be adapted for biomedical corpora. We introduce BioBERT (Bidirectional Encoder Representations from Transformers for Biomedical Text Mining), which is a domain-specific language representation model pre-trained on large-scale biomedical corpora. With almost the same architecture across tasks, BioBERT largely outperforms BERT and previous state-of-the-art models in a variety of biomedical text mining tasks when pre-trained on biomedical corpora. While BERT obtains performance comparable to that of previous state-of-the-art models, BioBERT significantly outperforms them on the following three representative biomedical text mining tasks: biomedical named entity recognition (0.62% F1 score improvement), biomedical relation extraction (2.80% F1 score improvement) and biomedical question answering (12.24% MRR improvement). Our analysis results show that pre-training BERT on biomedical corpora helps it to understand complex biomedical texts. We make the pre-trained weights of BioBERT freely available at https://github.com/naver/biobert-pretrained, and the source code for fine-tuning BioBERT available at https://github.com/dmis-lab/biobert.

BioBERT: A Pre-trained Biomedical Language Representation Model for Biomedical Text Mining

The paper "BioBERT: a pre-trained biomedical language representation model for biomedical text mining" introduces BioBERT, a language representation model specifically pre-trained on biomedical corpora to enhance the performance of various biomedical text mining tasks. Standard state-of-the-art NLP methodologies like BERT often exhibit sub-optimal results when directly applied to biomedical texts due to the unique terminologies and distributions inherent in biomedical corpora. This paper investigates the adaptation and pre-training of BERT on domain-specific corpora to bridge these gaps.

Motivation

The rapid accumulation of biomedical literature has spurred the necessity for sophisticated text mining tools to extract meaningful insights. Traditional NLP models pre-trained on general-domain corpora like Wikipedia and BooksCorpus face significant performance degradation when applied to biomedical text because of domain-specific terminology. This word distribution shift necessitates the development of domain-specific models.

Approach

BioBERT is conceptualized by initializing weights from the original BERT model, which was trained on general-domain corpora, and subsequently pre-training it on large-scale biomedical corpora, including PubMed abstracts and PMC full-text articles. The pre-training involves 23 days of computational effort utilizing eight NVIDIA V100 GPUs to absorb and generalize the biomedical domain's unique linguistic features. The fine-tuning of BioBERT is meticulously executed on three major biomedical text mining tasks: named entity recognition (NER), relation extraction (RE), and question answering (QA).

Results

BioBERT significantly outperforms both traditional models and the original BERT in various biomedical text mining tasks. Specific enhancements were observed in:

  • Biomedical Named Entity Recognition (NER): BioBERT achieved a 0.62% improvement in the F1 score.
  • Biomedical Relation Extraction (RE): It demonstrated a 2.80% improvement in the F1 score.
  • Biomedical Question Answering (QA): Here, BioBERT improved MRR scores by a substantial 12.24%.

These results underscore the proficiency of BioBERT in understanding complex biomedical texts and its superior performance over general-domain models.

Implications

The implications of the research are multifaceted. Practically, the deployment of BioBERT promises to enhance the efficiency and accuracy of biomedical information extraction systems, potentially accelerating research and discovery in the biomedical field. Theoretical implications suggest that domain-specific pre-training is a crucial factor in optimizing LLMs for specialized domains. Furthermore, the public release of BioBERT’s pre-trained weights and source code for fine-tuning opens avenues for future research and model enhancement by the wider scientific community.

Future Developments

BioBERT's success signifies a promising trajectory for domain-specific LLMs. Future developments could include:

  • Expansion into other specialized fields (e.g., legal or financial text mining).
  • Enhancing the model with additional biomedical corpora to further improve its domain adaptation.
  • Investigating the integration of BioBERT with other advanced NLP techniques such as reinforcement learning.

In conclusion, BioBERT represents a significant advancement in the application of deep learning models to the biomedical domain, offering substantial improvements in the critical tasks of NER, RE, and QA. This paper's release of BioBERT's pre-trained weights and fine-tuning resources facilitates ongoing improvements and applications by the biomedical research community.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Jinhyuk Lee (27 papers)
  2. Wonjin Yoon (13 papers)
  3. Sungdong Kim (30 papers)
  4. Donghyeon Kim (26 papers)
  5. Sunkyu Kim (3 papers)
  6. Chan Ho So (3 papers)
  7. Jaewoo Kang (83 papers)
Citations (4,994)
Youtube Logo Streamline Icon: https://streamlinehq.com