Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Class-Adaptive Self-Training for Relation Extraction with Incompletely Annotated Training Data (2306.09697v1)

Published 16 Jun 2023 in cs.CL

Abstract: Relation extraction (RE) aims to extract relations from sentences and documents. Existing relation extraction models typically rely on supervised machine learning. However, recent studies showed that many RE datasets are incompletely annotated. This is known as the false negative problem in which valid relations are falsely annotated as 'no_relation'. Models trained with such data inevitably make similar mistakes during the inference stage. Self-training has been proven effective in alleviating the false negative problem. However, traditional self-training is vulnerable to confirmation bias and exhibits poor performance in minority classes. To overcome this limitation, we proposed a novel class-adaptive re-sampling self-training framework. Specifically, we re-sampled the pseudo-labels for each class by precision and recall scores. Our re-sampling strategy favored the pseudo-labels of classes with high precision and low recall, which improved the overall recall without significantly compromising precision. We conducted experiments on document-level and biomedical relation extraction datasets, and the results showed that our proposed self-training framework consistently outperforms existing competitive methods on the Re-DocRED and ChemDisgene datasets when the training data are incompletely annotated. Our code is released at https://github.com/DAMO-NLP-SG/CAST.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (44)
  1. TACRED revisited: A thorough evaluation of the TACRED relation extraction task. In Proceedings of ACL.
  2. Pseudo-labeling and confirmation bias in deep semi-supervised learning. In Proceedings of IJCNN.
  3. H-FND: hierarchical false-negative denoising for distant supervision relation extraction. In Findings of ACL.
  4. A dataset for hyper-relational extraction and a cube-filling approach. In Proceedings of EMNLP.
  5. RelationPrompt: Leveraging prompts to generate synthetic data for zero-shot relation triplet extraction. In Findings of ACL.
  6. Junghoo Cho and Sourashis Roy. 2004. Impact of search engines on page popularity. In Proceedings of WWW, page 20–29.
  7. Connecting the dots: Document-level neural relation extraction with edge-oriented graphs. In Proceedings of EMNLP.
  8. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of NAACL.
  9. Semi-supervised classification for extracting protein interaction sentences using dependency parsing. In Proceedings of EMNLP-CoNLL.
  10. Reinforcement learning for relation classification from noisy data. In Proceedings of AAAI.
  11. Domain-specific language model pretraining for biomedical natural language processing. ACM Transactions on Computing for Healthcare (HEALTH).
  12. Towards integration of discriminability and robustness for document-level relation extraction. In Proceedings of EACL.
  13. Rethinking re-sampling in imbalanced semi-supervised learning. arXiv preprint arXiv:2106.00209.
  14. Semi-supervised relation extraction via incremental meta self-training. In Findings of EMNLP.
  15. Does recommend-revise produce reliable annotations? an analysis on missing instances in DocRED. In Proceedings of ACL.
  16. Better modeling of incomplete annotations for named entity recognition. In Proceedings of NAACL.
  17. Dividemix: Learning with noisy labels as semi-supervised learning. In Proceedings of ICLR.
  18. Aspect term extraction with history attention and selective transformation. In Proceedings of IJCAI.
  19. Empirical analysis of unlabeled entity problem in named entity recognition. In Proceedings of ICLR.
  20. Syntax role for neural semantic role labeling. Computational Linguistics, 47(3).
  21. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
  22. Distantly supervised named entity recognition using positive-unlabeled learning. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.
  23. Re-tacred: addressing shortcomings of the tacred dataset. In Proceedings of AAAI.
  24. Semi-supervised relation extraction with large-scale word clustering. In Proceedings of ACL.
  25. Document-level relation extraction with adaptive focal loss and knowledge distillation. In Findings of ACL.
  26. Revisiting docred–addressing the overlooked false negative problem in relation extraction. In Proceedings of EMNLP.
  27. Antti Tarvainen and Harri Valpola. 2017. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Proceedings of NIPS.
  28. Simultaneously self-attending to all mentions for full-abstract biological relation extraction. In Proceedings of NAACL.
  29. A unified positive-unlabeled learning framework for document-level relation extraction with different levels of labeling. In Proceedings of EMNLP.
  30. Tackling long-tailed relations and uncommon entities in knowledge graph completion. In Proceedings of EMNLP.
  31. Crest: A class-rebalancing self-training framework for imbalanced semi-supervised learning. In Proceedings of CVPR.
  32. Denoising relation extraction from document-level distant supervision. In Proceedings EMNLP.
  33. Better sampling of negatives for distantly supervised named entity recognition. In Findings of ACL.
  34. Better feature integration for named entity recognition. In Proceedings of NAACL.
  35. LUKE: Deep contextualized entity representations with entity-aware self-attention. In Proceedings of EMNLP.
  36. DocRED: a large-scale document-level relation extraction dataset. In Proceedings of ACL.
  37. Double graph based reasoning for document-level relation extraction. In Proceedings of EMNLP, Online.
  38. A distant supervision corpus for extracting biomedical relationships between chemicals, diseases and genes. In Proceedings of LREC.
  39. Document-level relation extraction as semantic segmentation. In Proceedings of IJCAI.
  40. Aspect sentiment quad prediction as paraphrase generation. In Proceedings of EMNLP.
  41. Position-aware attention and supervised data improve slot filling. In Proceedings of EMNLP, pages 35–45.
  42. Improving self-training for cross-lingual named entity recognition with contrastive and prototype learning. In Proceedings of ACL.
  43. Document-level relation extraction with adaptive thresholding and localized context pooling. In Proceedings of AAAI.
  44. Xiaojin Zhu and Andrew B. Goldberg. 2009. Introduction to semi-supervised learning. In Introduction to Semi-Supervised Learning.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Qingyu Tan (9 papers)
  2. Lu Xu (68 papers)
  3. Lidong Bing (144 papers)
  4. Hwee Tou Ng (44 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.