On the use of Silver Standard Data for Zero-shot Classification Tasks in Information Extraction (2402.18061v2)
Abstract: The superior performance of supervised classification methods in the information extraction (IE) area heavily relies on a large amount of gold standard data. Recent zero-shot classification methods converted the task to other NLP tasks (e.g., textual entailment) and used off-the-shelf models of these NLP tasks to directly perform inference on the test data without using a large amount of IE annotation data. A potentially valuable by-product of these methods is the large-scale silver standard data, i.e., pseudo-labeled data by the off-the-shelf models of other NLP tasks. However, there is no further investigation into the use of these data. In this paper, we propose a new framework, Clean-LaVe, which aims to utilize silver standard data to enhance the zero-shot performance. Clean-LaVe includes four phases: (1) Obtaining silver data; (2) Identifying relatively clean data from silver data; (3) Finetuning the off-the-shelf model using clean data; (4) Inference on the test data. The experimental results show that Clean-LaVe can outperform the baseline by 5% and 6% on TACRED and Wiki80 dataset in the zero-shot relation classification task, and by 3%-7% on Smile (Korean and Polish) in the zero-shot cross-lingual relation classification task, and by 8% on ACE05-E+ in the zero-shot event argument classification task. The code is share in https://github.com/wjw136/Clean_LaVe.git.
- Open AI. 2023. Gpt-4 technical report. arXiv preprint arXiv:2303.08774.
- Görkem Algan and Ilkay Ulusoy. 2021. Image classification with deep learning in the presence of noisy labels: A survey. Knowledge Based System, 215:106771.
- Unsupervised label noise modeling and loss correction. In Proceedings of ICML, pages 312–321.
- A closer look at memorization in deep networks. In Proceedings of ICML, pages 233–242.
- On symmetric losses for learning from corrupted labels. In Proceedings of ICML, pages 961–970.
- Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of NAACL-HLT, pages 4171–4186.
- Benoît Frénay and Michel Verleysen. 2013. Classification in the presence of label noise: a survey. IEEE transactions on neural networks and learning systems, 25(5):845–869.
- Robust loss functions under label noise for deep neural networks. In Proceedings of the AAAI, pages 1919–1925.
- Unsupervised relation extraction from language models using constrained cloze completion. In Findings of EMNLP, pages 1263–1276.
- A survey of label-noise representation learning: Past, present and future. arXiv preprint.
- Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Proceedings of NeurIPS, pages 8536–8546.
- Deberta: Decoding-enhanced bert with disentangled attention. In Proceedings of ICLR.
- O2u-net: A simple noisy label detection approach for deep neural networks. In Proceedings of ICCV, pages 3326–3334.
- Multilingual generative language models for zero-shot cross-lingual event argument extraction. In Proceedings ACL, pages 4633–4646.
- Zero-shot transfer learning for event extraction. In Proceedings of ACL, pages 2160–2170.
- Uncertainty-aware learning against label noise on imbalanced datasets. In Proceedings of AAAI, volume 36, pages 6960–6969.
- Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In Proceedings of ICML, pages 2304–2313.
- Instance-adaptive training with noise-robust losses against noisy labels. In Proceedings of EMNLP, pages 5647–5663.
- Nlnl: Negative learning for noisy labels. In Proceedings of ICCV, pages 101–110.
- Less Annotating, More Classifying – Addressing the Data Scarcity Issue of Supervised Machine Learning with Deep Transfer Learning and BERT - NLI. Preprint. Publisher: Open Science Framework.
- Zero-shot relation extraction via reading comprehension. In Proceedings of CoNLL, pages 333–342.
- Dividemix: Learning with noisy labels as semi-supervised learning. In Proceedings of ICLR.
- Global constraints with prompting for zero-shot event argument classification. In Findings of EACL, pages 2482–2493.
- Event extraction as machine reading comprehension. In Proceedings of EMNLP, pages 1641–1651.
- Translation-based implicit annotation projection for zero-shot cross-lingual event argument extraction. In Proceedings of SIGIR, pages 2076–2081.
- Event extraction as question generation and answering. In Proceedings of ACL, pages 1666–1688.
- Summarization as indirect supervision for relation extraction. arXiv preprint arXiv:2205.09837.
- Label-noise learning with intrinsically long-tailed data. In Proceedings of ICCV, pages 1369–1378.
- Zero-shot event extraction via transfer learning: Challenges and insights. In Proceedings of ACL, pages 322–332.
- Shengfei Lyu and Huanhuan Chen. 2021. Relation classification with entity type restriction. In Proceedings of ACL, pages 390–395.
- Yueming Lyu and Ivor W Tsang. 2019. Curriculum loss: Robust learning and generalization against label corruption. In Proceedings of ICLR.
- Wider & closer: Mixture of short-channel distillers for zero-shot cross-lingual named entity recognition. arXiv preprint arXiv:2212.03506.
- Shine: Syntax-augmented hierarchical interactive encoder for zero-shot cross-lingual information extraction. arXiv preprint arXiv:2305.12389.
- Normalized loss functions for deep learning with noisy labels. In Proceedings of ICML, volume 119, pages 6543–6553.
- Eran Malach and Shai Shalev-Shwartz. 2017. Decoupling" when to update" from" how to update". In Proceedings of NeurIPS, pages 960–970.
- Improving zero-shot event extraction via sentence simplification. In Workshop of EMNLP, pages 32–43.
- Can gradient clipping mitigate label noise? In Proceedings of ICLR.
- Abiola Obamuyide and Andreas Vlachos. 2018. Zero-shot relation classification as textual entailment. In Proceedings of the FEVER, pages 72–78.
- Sinno Jialin Pan and Qiang Yang. 2009. A survey on transfer learning. IEEE Transactions on knowledge and data engineering, pages 1345–1359.
- Mahdi Rahimi and Mihai Surdeanu. 2023. Improving zero-shot relation classification via automatically-acquired entailment templates. In Proceedings of Workshop on RepL4NLP, pages 187–195.
- Training deep neural networks on noisy labels with bootstrapping. In Proceedings of ICLR Workshop.
- Label verbalization and entailment for effective zero-and few-shot relation extraction. In Proceedings of EMNLP, pages 1199–1212.
- Textual entailment for event argument extraction: Zero- and few-shot with multi-source learning. In Findings of NAACL-HLT, pages 2439–2455.
- Zs4ie: A toolkit for zero-shot information extraction with simple verbalizations. In Proceedings of NAACL-HLT, pages 27–38.
- Parallel instance query network for named entity recognition. In Proceedings of ACL, pages 947–961.
- Meta-weight-net: Learning an explicit mapping for sample weighting. In Proceedings of NeurIPS, pages 1917–1928.
- Combating label noise in deep learning using abstention. arXiv preprint arXiv:1905.10964.
- Revisiting unsupervised relation extraction. In Proceedings of ACL, pages 7498–7505.
- One-shot to weakly-supervised relation classification using language models. In Proceedings of AKBC.
- Symmetric cross entropy for robust learning with noisy labels. In Proceedings of ICCV, pages 322–330.
- Single-/multi-source cross-lingual ner via teacher-student learning on unlabeled data in target language. In Proceedings of ACL, pages 6505–6514.
- Exploring pre-trained language models for event extraction and generation. In Proceedings of the 57th annual meeting of the association for computational linguistics, pages 5284–5294.
- Searching to exploit memorization effect in learning from corrupted labels. arXiv preprint arXiv:1911.02377.
- How does disagreement help generalization against label corruption? In Proceedings of ICML, pages 7164–7173.
- Understanding deep learning requires rethinking generalization. In Proceedings of ICLR.
- Understanding deep learning (still) requires rethinking generalization. Communications of the ACM, 64(3):107–115.
- Zero-shot label-aware event trigger and argument classification. In Findings of ACL-IJCNLP, pages 1331–1340.
- Aligning instruction tasks unlocks large language models as zero-shot relation extractors. In ACL, pages 794–812.
- Zhilu Zhang and Mert Sabuncu. 2018. Generalized cross entropy loss for training deep neural networks with noisy labels. In Proceedings of NeurIPS, pages 8792–8802.
- Re-matching: A fine-grained semantic matching method for zero-shot relation extraction. In Proceedings of ACL, pages 6680–6691.
- Zexuan Zhong and Danqi Chen. 2021. A frustratingly easy approach for entity and relation extraction. In Proceedings of NAACL, pages 50–61.
- Wenxuan Zhou and Muhao Chen. 2021. Learning from noisy labels for entity-centric information extraction. In Proceedings of EMNLP, pages 5381–5392.
- Enwei Zhu and Jinpeng Li. 2022. Boundary smoothing for named entity recognition. In Proceedings of ACL, pages 7096–7108.
- A joint neural model for information extraction with global features. PID https://aclanthology.org/2020.acl-main.713/.
- Multilingual Entity and Relation Extraction Dataset and Model. PID https://aclanthology.org/2021.eacl-main.166.
- Hongmin Xiao. 2022. Wiki80. PID https://figshare.com/articles/dataset/Wiki80/19323371.
- Victor Zhong and Yuhao Zhang and Danqi Chen and Gabor Angeli and Christopher Manning. 2018. TAC Relation Extraction Dataset. ISLRN 927-859-759-915-2. PID https://catalog.ldc.upenn.edu/LDC2018T24.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.