Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adaptive Self-training for Few-shot Neural Sequence Labeling (2010.03680v2)

Published 7 Oct 2020 in cs.CL, cs.AI, and cs.LG

Abstract: Sequence labeling is an important technique employed for many NLP tasks, such as Named Entity Recognition (NER), slot tagging for dialog systems and semantic parsing. Large-scale pre-trained LLMs obtain very good performance on these tasks when fine-tuned on large amounts of task-specific labeled data. However, such large-scale labeled datasets are difficult to obtain for several tasks and domains due to the high cost of human annotation as well as privacy and data access constraints for sensitive user applications. This is exacerbated for sequence labeling tasks requiring such annotations at token-level. In this work, we develop techniques to address the label scarcity challenge for neural sequence labeling models. Specifically, we develop self-training and meta-learning techniques for training neural sequence taggers with few labels. While self-training serves as an effective mechanism to learn from large amounts of unlabeled data -- meta-learning helps in adaptive sample re-weighting to mitigate error propagation from noisy pseudo-labels. Extensive experiments on six benchmark datasets including two for massive multilingual NER and four slot tagging datasets for task-oriented dialog systems demonstrate the effectiveness of our method. With only 10 labeled examples for each class for each task, our method obtains 10% improvement over state-of-the-art systems demonstrating its effectiveness for the low-resource setting.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Yaqing Wang (59 papers)
  2. Subhabrata Mukherjee (59 papers)
  3. Haoda Chu (2 papers)
  4. Yuancheng Tu (2 papers)
  5. Ming Wu (43 papers)
  6. Jing Gao (98 papers)
  7. Ahmed Hassan Awadallah (50 papers)
Citations (33)