In-Context Learning for Few-Shot Nested Named Entity Recognition (2402.01182v1)
Abstract: In nested Named entity recognition (NER), entities are nested with each other, and thus requiring more data annotations to address. This leads to the development of few-shot nested NER, where the prevalence of pretrained LLMs with in-context learning (ICL) offers promising solutions. In this work, we introduce an effective and innovative ICL framework for the setting of few-shot nested NER. We improve the ICL prompt by devising a novel example demonstration selection mechanism, EnDe retriever. In EnDe retriever, we employ contrastive learning to perform three types of representation learning, in terms of semantic similarity, boundary similarity, and label similarity, to generate high-quality demonstration examples. Extensive experiments over three nested NER and four flat NER datasets demonstrate the efficacy of our system.
- Erik F. Tjong Kim Sang and Fien De Meulder, “Introduction to the CoNLL-2003 shared task: Language-independent named entity recognition,” in Proceedings of NAACL, 2003, pp. 142–147.
- “Nested named entity recognition,” in Proceedings of EMNLP, 2009, pp. 141–150.
- “Boundaries and edges rethinking: An end-to-end neural model for overlapping entity relation extraction,” Information Processing & Management, vol. 57, no. 6, pp. 102311, 2020.
- “A neural layered model for nested named entity recognition,” in Proceedings of NAACL, 2018, pp. 1446–1459.
- “Joint mention extraction and classification with mention hypergraphs,” in Proceedings of EMNLP, 2015, pp. 857–867.
- “A unified generative framework for various NER subtasks,” in Proceedings of ACL, 2021, pp. 5808–5822.
- “Lasuie: Unifying information extraction with latent adaptive structure-aware generative language model,” in Proceedings of NeurIPS, 2022.
- “Rethinking boundaries: End-to-end recognition of discontinuous mentions with pointer networks,” in Proceedings of AAAI, 2021, pp. 12785–12793.
- “Exploring the limits of transfer learning with a unified text-to-text transformer,” Journal of Machine Learning Research, vol. 21, no. 140, pp. 1–67, 2020.
- “Next-gpt: Any-to-any multimodal llm,” 2023.
- “Few-shot named entity recognition: A comprehensive study,” CoRR, vol. abs/2012.14978, 2020.
- “Few-shot named entity recognition with self-describing networks,” in Proceedings of ACL, 2022, pp. 5711–5722.
- “Language models are few-shot learners,” in NeurIPS, 2020.
- “Rethinking the role of demonstrations: What makes in-context learning work?,” in Proceedings of EMNLP, 2022, pp. 11048–11064.
- “Better combine them together! integrating syntactic constituency and dependency representations for semantic role labeling,” in Findings of the ACL, 2021, pp. 549–559.
- “Learning a similarity metric discriminatively, with application to face verification,” in Proceedings of CVPR, 2005, vol. 1, pp. 539–546.
- “Locate and label: A two-stage identifier for nested named entity recognition,” in Proceedings of ACL, 2021, pp. 2782–2794.
- “SEE-few: Seed, expand and entail for few-shot named entity recognition,” in Proceedings of COLING, 2022, pp. 2540–2550.
- “An enhanced span-based decomposition method for few-shot sequence labeling,” in Proceedings of NAACL, 2022, pp. 5012–5024.
- “Unified named entity recognition as word-word relation classification,” in Proceedings of AAAI, 2022, pp. 10965–10973.
- “Focusing, bridging and prompting for few-shot nested named entity recognition,” in Findings of ACL, 2023, pp. 2621–2637.
- “BERT: Pre-training of deep bidirectional transformers for language understanding,” in Proceedings of NAACL, 2019, pp. 4171–4186.
- Meishan Zhang (70 papers)
- Bin Wang (750 papers)
- Hao Fei (105 papers)
- Min Zhang (630 papers)