Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Self-Improving for Zero-Shot Named Entity Recognition with Large Language Models (2311.08921v3)

Published 15 Nov 2023 in cs.CL
Self-Improving for Zero-Shot Named Entity Recognition with Large Language Models

Abstract: Exploring the application of powerful LLMs on the named entity recognition (NER) task has drawn much attention recently. This work pushes the performance boundary of zero-shot NER with LLMs by proposing a training-free self-improving framework, which utilizes an unlabeled corpus to stimulate the self-learning ability of LLMs. First, we use the LLM to make predictions on the unlabeled corpus using self-consistency and obtain a self-annotated dataset. Second, we explore various strategies to select reliable annotations to form a reliable self-annotated dataset. Finally, for each test input, we retrieve demonstrations from the reliable self-annotated dataset and perform inference via in-context learning. Experiments on four benchmarks show substantial performance improvements achieved by our framework. Through comprehensive experimental analysis, we find that increasing the size of unlabeled corpus or iterations of self-improving does not guarantee further improvement, but the performance might be boosted via more advanced strategies for reliable annotation selection. Code and data are publicly available at https://github.com/Emma1066/Self-Improve-Zero-Shot-NER

Summary of "Self-Improving for Zero-Shot Named Entity Recognition with LLMs"

The paper explores a novel, training-free framework to enhance zero-shot Named Entity Recognition (NER) capabilities using LLMs. By leveraging the self-learning abilities of LLMs in a resource-constrained setting, the authors propose a self-improving paradigm that significantly boosts performance over existing zero-shot approaches.

Framework Overview

The proposed framework operates under strict zero-shot conditions, where no annotated data or auxiliary models are available, and consists of three main steps:

  1. Zero-Shot Self-Annotating:
    • The framework uses LLMs to generate predictions on an unlabeled corpus. The process employs self-consistency (SC) to produce a self-annotated dataset, with each entity prediction accompanied by a self-consistency score that indicates the reliability of the annotation.
  2. Reliable Annotation Selection:
    • Several strategies are employed to filter the self-annotated data, enhancing annotation reliability. These include:
      • Entity-level threshold filtering: Discards entity annotations with SC scores below a pre-defined threshold.
      • Sample-level threshold filtering: Eliminates entire samples if their average SC score falls below a specified benchmark.
      • Two-stage majority voting: Selects the most consistent entities and their corresponding types.
  3. Inference with Self-Annotated Demonstrations:
    • For each test input, the framework retrieves demonstrations from the reliable self-annotated dataset, facilitating inference through in-context learning. Various retrieval strategies, including nearest neighbor and diverse sampling methods, are analyzed in the context of their performance.

Results and Implications

Experimental evaluation on four benchmark datasets reveals that this framework provides substantial performance gains compared to baseline approaches that do not utilize self-improving techniques. Importantly, the paper finds that simply increasing the size of the unlabeled corpus or iterations of self-improvement doesn't guarantee further performance optimization. Instead, improvements might be achievable through more sophisticated methods for selecting reliable annotations.

Technical Insights and Implications

  • The reliability of the self-annotation is crucial, underscored by the analysis of SC scores and the proposal of varied filtering thresholds.
  • The diverse nearest SC ranking method introduced by the authors strikes a better balance between similarity, diversity, and annotation reliability.
  • When comparing against available gold-labeled data, the gap highlights areas where future advancements could seek to refine self-improvement strategies more effectively.

Future Directions

The paper suggests several avenues for future research:

  • Exploration of advanced annotation selection strategies to improve the quality of self-labeled data.
  • Extending the self-improving paradigm to other information extraction tasks and beyond lexical entities.
  • Investigating the utility of additional LLM capabilities, such as self-verification, to mitigate inherent annotation noise.

Overall, the paper introduces an innovative paradigm for zero-shot NER, enhancing LLM self-improvement capabilities without the requirement for extensive computational resources or large-scale training datasets. The findings underscore the potential of LLMs in zero-shot settings while providing a foundation for more refined self-improving techniques.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Tingyu Xie (6 papers)
  2. Qi Li (352 papers)
  3. Yan Zhang (954 papers)
  4. Zuozhu Liu (78 papers)
  5. Hongwei Wang (150 papers)
Citations (12)
Youtube Logo Streamline Icon: https://streamlinehq.com