Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

NER-BERT: A Pre-trained Model for Low-Resource Entity Tagging (2112.00405v1)

Published 1 Dec 2021 in cs.CL and cs.AI

Abstract: Named entity recognition (NER) models generally perform poorly when large training datasets are unavailable for low-resource domains. Recently, pre-training a large-scale LLM has become a promising direction for coping with the data scarcity issue. However, the underlying discrepancies between the LLMing and NER task could limit the models' performance, and pre-training for the NER task has rarely been studied since the collected NER datasets are generally small or large but with low quality. In this paper, we construct a massive NER corpus with a relatively high quality, and we pre-train a NER-BERT model based on the created dataset. Experimental results show that our pre-trained model can significantly outperform BERT as well as other strong baselines in low-resource scenarios across nine diverse domains. Moreover, a visualization of entity representations further indicates the effectiveness of NER-BERT for categorizing a variety of entities.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Zihan Liu (102 papers)
  2. Feijun Jiang (13 papers)
  3. Yuxiang Hu (25 papers)
  4. Chen Shi (55 papers)
  5. Pascale Fung (151 papers)
Citations (34)

Summary

We haven't generated a summary for this paper yet.