Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Probing Biomedical Embeddings from Language Models (1904.02181v1)

Published 3 Apr 2019 in cs.CL

Abstract: Contextualized word embeddings derived from pre-trained LLMs (LMs) show significant improvements on downstream NLP tasks. Pre-training on domain-specific corpora, such as biomedical articles, further improves their performance. In this paper, we conduct probing experiments to determine what additional information is carried intrinsically by the in-domain trained contextualized embeddings. For this we use the pre-trained LMs as fixed feature extractors and restrict the downstream task models to not have additional sequence modeling layers. We compare BERT, ELMo, BioBERT and BioELMo, a biomedical version of ELMo trained on 10M PubMed abstracts. Surprisingly, while fine-tuned BioBERT is better than BioELMo in biomedical NER and NLI tasks, as a fixed feature extractor BioELMo outperforms BioBERT in our probing tasks. We use visualization and nearest neighbor analysis to show that better encoding of entity-type and relational information leads to this superiority.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Qiao Jin (74 papers)
  2. Bhuwan Dhingra (66 papers)
  3. William W. Cohen (79 papers)
  4. Xinghua Lu (16 papers)
Citations (116)
X Twitter Logo Streamline Icon: https://streamlinehq.com