Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multi-task Transformer with Relation-attention and Type-attention for Named Entity Recognition (2303.10870v1)

Published 20 Mar 2023 in cs.CL

Abstract: Named entity recognition (NER) is an important research problem in natural language processing. There are three types of NER tasks, including flat, nested and discontinuous entity recognition. Most previous sequential labeling models are task-specific, while recent years have witnessed the rising of generative models due to the advantage of unifying all NER tasks into the seq2seq model framework. Although achieving promising performance, our pilot studies demonstrate that existing generative models are ineffective at detecting entity boundaries and estimating entity types. This paper proposes a multi-task Transformer, which incorporates an entity boundary detection task into the named entity recognition task. More concretely, we achieve entity boundary detection by classifying the relations between tokens within the sentence. To improve the accuracy of entity-type mapping during decoding, we adopt an external knowledge base to calculate the prior entity-type distributions and then incorporate the information into the model via the self and cross-attention mechanisms. We perform experiments on an extensive set of NER benchmarks, including two flat, three nested, and three discontinuous NER datasets. Experimental results show that our approach considerably improves the generative NER model's performance.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Ying Mo (5 papers)
  2. Hongyin Tang (9 papers)
  3. Jiahao Liu (72 papers)
  4. Qifan Wang (129 papers)
  5. Zenglin Xu (145 papers)
  6. Jingang Wang (71 papers)
  7. Wei Wu (481 papers)
  8. Zhoujun Li (122 papers)
Citations (4)