Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

NER-BERT: A Pre-trained Model for Low-Resource Entity Tagging (2112.00405v1)

Published 1 Dec 2021 in cs.CL and cs.AI

Abstract: Named entity recognition (NER) models generally perform poorly when large training datasets are unavailable for low-resource domains. Recently, pre-training a large-scale LLM has become a promising direction for coping with the data scarcity issue. However, the underlying discrepancies between the LLMing and NER task could limit the models' performance, and pre-training for the NER task has rarely been studied since the collected NER datasets are generally small or large but with low quality. In this paper, we construct a massive NER corpus with a relatively high quality, and we pre-train a NER-BERT model based on the created dataset. Experimental results show that our pre-trained model can significantly outperform BERT as well as other strong baselines in low-resource scenarios across nine diverse domains. Moreover, a visualization of entity representations further indicates the effectiveness of NER-BERT for categorizing a variety of entities.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Zihan Liu (102 papers)
  2. Feijun Jiang (13 papers)
  3. Yuxiang Hu (25 papers)
  4. Chen Shi (55 papers)
  5. Pascale Fung (150 papers)
Citations (34)