Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Application of Pre-training Models in Named Entity Recognition (2002.08902v1)

Published 9 Feb 2020 in cs.CL, cs.LG, and stat.ML

Abstract: Named Entity Recognition (NER) is a fundamental NLP task to extract entities from unstructured data. The previous methods for NER were based on machine learning or deep learning. Recently, pre-training models have significantly improved performance on multiple NLP tasks. In this paper, firstly, we introduce the architecture and pre-training tasks of four common pre-training models: BERT, ERNIE, ERNIE2.0-tiny, and RoBERTa. Then, we apply these pre-training models to a NER task by fine-tuning, and compare the effects of the different model architecture and pre-training tasks on the NER task. The experiment results showed that RoBERTa achieved state-of-the-art results on the MSRA-2006 dataset.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yu Wang (939 papers)
  2. Yining Sun (8 papers)
  3. Zuchang Ma (1 paper)
  4. Lisheng Gao (2 papers)
  5. Yang Xu (277 papers)
  6. Ting Sun (26 papers)
Citations (20)