Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

In-Context Learning for Few-Shot Nested Named Entity Recognition (2402.01182v1)

Published 2 Feb 2024 in cs.CL

Abstract: In nested Named entity recognition (NER), entities are nested with each other, and thus requiring more data annotations to address. This leads to the development of few-shot nested NER, where the prevalence of pretrained LLMs with in-context learning (ICL) offers promising solutions. In this work, we introduce an effective and innovative ICL framework for the setting of few-shot nested NER. We improve the ICL prompt by devising a novel example demonstration selection mechanism, EnDe retriever. In EnDe retriever, we employ contrastive learning to perform three types of representation learning, in terms of semantic similarity, boundary similarity, and label similarity, to generate high-quality demonstration examples. Extensive experiments over three nested NER and four flat NER datasets demonstrate the efficacy of our system.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. Erik F. Tjong Kim Sang and Fien De Meulder, “Introduction to the CoNLL-2003 shared task: Language-independent named entity recognition,” in Proceedings of NAACL, 2003, pp. 142–147.
  2. “Nested named entity recognition,” in Proceedings of EMNLP, 2009, pp. 141–150.
  3. “Boundaries and edges rethinking: An end-to-end neural model for overlapping entity relation extraction,” Information Processing & Management, vol. 57, no. 6, pp. 102311, 2020.
  4. “A neural layered model for nested named entity recognition,” in Proceedings of NAACL, 2018, pp. 1446–1459.
  5. “Joint mention extraction and classification with mention hypergraphs,” in Proceedings of EMNLP, 2015, pp. 857–867.
  6. “A unified generative framework for various NER subtasks,” in Proceedings of ACL, 2021, pp. 5808–5822.
  7. “Lasuie: Unifying information extraction with latent adaptive structure-aware generative language model,” in Proceedings of NeurIPS, 2022.
  8. “Rethinking boundaries: End-to-end recognition of discontinuous mentions with pointer networks,” in Proceedings of AAAI, 2021, pp. 12785–12793.
  9. “Exploring the limits of transfer learning with a unified text-to-text transformer,” Journal of Machine Learning Research, vol. 21, no. 140, pp. 1–67, 2020.
  10. “Next-gpt: Any-to-any multimodal llm,” 2023.
  11. “Few-shot named entity recognition: A comprehensive study,” CoRR, vol. abs/2012.14978, 2020.
  12. “Few-shot named entity recognition with self-describing networks,” in Proceedings of ACL, 2022, pp. 5711–5722.
  13. “Language models are few-shot learners,” in NeurIPS, 2020.
  14. “Rethinking the role of demonstrations: What makes in-context learning work?,” in Proceedings of EMNLP, 2022, pp. 11048–11064.
  15. “Better combine them together! integrating syntactic constituency and dependency representations for semantic role labeling,” in Findings of the ACL, 2021, pp. 549–559.
  16. “Learning a similarity metric discriminatively, with application to face verification,” in Proceedings of CVPR, 2005, vol. 1, pp. 539–546.
  17. “Locate and label: A two-stage identifier for nested named entity recognition,” in Proceedings of ACL, 2021, pp. 2782–2794.
  18. “SEE-few: Seed, expand and entail for few-shot named entity recognition,” in Proceedings of COLING, 2022, pp. 2540–2550.
  19. “An enhanced span-based decomposition method for few-shot sequence labeling,” in Proceedings of NAACL, 2022, pp. 5012–5024.
  20. “Unified named entity recognition as word-word relation classification,” in Proceedings of AAAI, 2022, pp. 10965–10973.
  21. “Focusing, bridging and prompting for few-shot nested named entity recognition,” in Findings of ACL, 2023, pp. 2621–2637.
  22. “BERT: Pre-training of deep bidirectional transformers for language understanding,” in Proceedings of NAACL, 2019, pp. 4171–4186.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Meishan Zhang (70 papers)
  2. Bin Wang (750 papers)
  3. Hao Fei (105 papers)
  4. Min Zhang (630 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets