Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Pre-trained Language Model for Biomedical Question Answering (1909.08229v1)

Published 18 Sep 2019 in cs.CL

Abstract: The recent success of question answering systems is largely attributed to pre-trained LLMs. However, as LLMs are mostly pre-trained on general domain corpora such as Wikipedia, they often have difficulty in understanding biomedical questions. In this paper, we investigate the performance of BioBERT, a pre-trained biomedical LLM, in answering biomedical questions including factoid, list, and yes/no type questions. BioBERT uses almost the same structure across various question types and achieved the best performance in the 7th BioASQ Challenge (Task 7b, Phase B). BioBERT pre-trained on SQuAD or SQuAD 2.0 easily outperformed previous state-of-the-art models. BioBERT obtains the best performance when it uses the appropriate pre-/post-processing strategies for questions, passages, and answers.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Wonjin Yoon (13 papers)
  2. Jinhyuk Lee (27 papers)
  3. Donghyeon Kim (26 papers)
  4. Minbyul Jeong (18 papers)
  5. Jaewoo Kang (83 papers)
Citations (84)