Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning In-context Learning for Named Entity Recognition (2305.11038v3)

Published 18 May 2023 in cs.CL

Abstract: Named entity recognition in real-world applications suffers from the diversity of entity types, the emergence of new entity types, and the lack of high-quality annotations. To address the above problems, this paper proposes an in-context learning-based NER approach, which can effectively inject in-context NER ability into PLMs and recognize entities of novel types on-the-fly using only a few demonstrative instances. Specifically, we model PLMs as a meta-function $\mathcal{ \lambda_ {\text{instruction, demonstrations, text}}. M}$, and a new entity extractor can be implicitly constructed by applying new instruction and demonstrations to PLMs, i.e., $\mathcal{ (\lambda . M) }$(instruction, demonstrations) $\to$ $\mathcal{F}$ where $\mathcal{F}$ will be a new entity extractor, i.e., $\mathcal{F}$: text $\to$ entities. To inject the above in-context NER ability into PLMs, we propose a meta-function pre-training algorithm, which pre-trains PLMs by comparing the (instruction, demonstration)-initialized extractor with a surrogate golden extractor. Experimental results on 4 few-shot NER datasets show that our method can effectively inject in-context NER ability into PLMs and significantly outperforms the PLMs+fine-tuning counterparts.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Jiawei Chen (161 papers)
  2. Yaojie Lu (61 papers)
  3. Hongyu Lin (94 papers)
  4. Jie Lou (34 papers)
  5. Wei Jia (52 papers)
  6. Dai Dai (3 papers)
  7. Hua Wu (191 papers)
  8. Boxi Cao (21 papers)
  9. Xianpei Han (103 papers)
  10. Le Sun (111 papers)
Citations (13)

Summary

We haven't generated a summary for this paper yet.