Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Contextual embedding and model weighting by fusing domain knowledge on Biomedical Question Answering (2206.12866v1)

Published 26 Jun 2022 in cs.CL and cs.AI

Abstract: Biomedical Question Answering aims to obtain an answer to the given question from the biomedical domain. Due to its high requirement of biomedical domain knowledge, it is difficult for the model to learn domain knowledge from limited training data. We propose a contextual embedding method that combines open-domain QA model \aoa and \biobert model pre-trained on biomedical domain data. We adopt unsupervised pre-training on large biomedical corpus and supervised fine-tuning on biomedical question answering dataset. Additionally, we adopt an MLP-based model weighting layer to automatically exploit the advantages of two models to provide the correct answer. The public dataset \biomrc constructed from PubMed corpus is used to evaluate our method. Experimental results show that our model outperforms state-of-the-art system by a large margin.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Yuxuan Lu (26 papers)
  2. Jingya Yan (2 papers)
  3. Zhixuan Qi (3 papers)
  4. Zhongzheng Ge (1 paper)
  5. Yongping Du (21 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.