Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Enhancing Clinical Information Extraction with Transferred Contextual Embeddings (2109.07243v2)

Published 15 Sep 2021 in cs.CL

Abstract: The Bidirectional Encoder Representations from Transformers (BERT) model has achieved the state-of-the-art performance for many NLP tasks. Yet, limited research has been contributed to studying its effectiveness when the target domain is shifted from the pre-training corpora, for example, for biomedical or clinical NLP applications. In this paper, we applied it to a widely studied a hospital information extraction (IE) task and analyzed its performance under the transfer learning setting. Our application became the new state-of-the-art result by a clear margin, compared with a range of existing IE models. Specifically, on this nursing handover data set, the macro-average F1 score from our model was 0.438, whilst the previous best deep learning models had 0.416. In conclusion, we showed that BERT based pre-training models can be transferred to health-related documents under mild conditions and with a proper fine-tuning process.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Zimin Wan (2 papers)
  2. Chenchen Xu (12 papers)
  3. Hanna Suominen (17 papers)