Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Robust Self-Augmentation for Named Entity Recognition with Meta Reweighting (2204.11406v4)

Published 25 Apr 2022 in cs.CL and cs.AI

Abstract: Self-augmentation has received increasing research interest recently to improve named entity recognition (NER) performance in low-resource scenarios. Token substitution and mixup are two feasible heterogeneous self-augmentation techniques for NER that can achieve effective performance with certain specialized efforts. Noticeably, self-augmentation may introduce potentially noisy augmented data. Prior research has mainly resorted to heuristic rule-based constraints to reduce the noise for specific self-augmentation methods individually. In this paper, we revisit these two typical self-augmentation methods for NER, and propose a unified meta-reweighting strategy for them to achieve a natural integration. Our method is easily extensible, imposing little effort on a specific self-augmentation method. Experiments on different Chinese and English NER benchmarks show that our token substitution and mixup method, as well as their integration, can achieve effective performance improvement. Based on the meta-reweighting mechanism, we can enhance the advantages of the self-augmentation techniques without much extra effort.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Linzhi Wu (6 papers)
  2. Pengjun Xie (85 papers)
  3. Jie Zhou (687 papers)
  4. Meishan Zhang (70 papers)
  5. Chunping Ma (1 paper)
  6. Guangwei Xu (18 papers)
  7. Min Zhang (630 papers)
Citations (16)
Github Logo Streamline Icon: https://streamlinehq.com