Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 85 tok/s
Gemini 2.5 Pro 36 tok/s Pro
GPT-5 Medium 19 tok/s Pro
GPT-5 High 20 tok/s Pro
GPT-4o 72 tok/s Pro
Kimi K2 170 tok/s Pro
GPT OSS 120B 457 tok/s Pro
Claude Sonnet 4 38 tok/s Pro
2000 character limit reached

Information Extraction in Low-Resource Scenarios: Survey and Perspective (2202.08063v6)

Published 16 Feb 2022 in cs.CL, cs.AI, cs.IR, and cs.LG

Abstract: Information Extraction (IE) seeks to derive structured information from unstructured texts, often facing challenges in low-resource scenarios due to data scarcity and unseen classes. This paper presents a review of neural approaches to low-resource IE from \emph{traditional} and \emph{LLM-based} perspectives, systematically categorizing them into a fine-grained taxonomy. Then we conduct empirical study on LLM-based methods compared with previous state-of-the-art models, and discover that (1) well-tuned LMs are still predominant; (2) tuning open-resource LLMs and ICL with GPT family is promising in general; (3) the optimal LLM-based technical solution for low-resource IE can be task-dependent. In addition, we discuss low-resource IE with LLMs, highlight promising applications, and outline potential research directions. This survey aims to foster understanding of this field, inspire new ideas, and encourage widespread applications in both academia and industry.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (158)
  1. Adapting pre-trained language models to african languages via multilingual adaptive fine-tuning. In COLING, pages 4336–4349. International Committee on Computational Linguistics.
  2. TACRED revisited: A thorough evaluation of the TACRED relation extraction task. In ACL, pages 1558–1569. Association for Computational Linguistics.
  3. Language models are few-shot learners. In NeurIPS.
  4. Improving low-resource named entity recognition with graph propagated data augmentation. In ACL (Short), pages 110––118. Association for Computational Linguistics.
  5. Joint multimedia event extraction from video and article. In EMNLP (Findings), pages 74–88. Association for Computational Linguistics.
  6. Chih-Yao Chen and Cheng-Te Li. 2021. ZS-BERT: towards zero-shot relation extraction with attribute representation learning. In NAACL-HLT, pages 3470–3479. Association for Computational Linguistics.
  7. Local additivity based data augmentation for semi-supervised NER. In EMNLP (1), pages 1241–1251. Association for Computational Linguistics.
  8. Few-shot named entity recognition with self-describing networks. In ACL (1), pages 5711–5722. Association for Computational Linguistics.
  9. Learning in-context learning for named entity recognition. In ACL (1), pages 13661–13675. Association for Computational Linguistics.
  10. Evaluating large language models trained on code. CoRR, abs/2107.03374.
  11. Heproto: A hierarchical enhancing protonet based on multi-task learning for few-shot named entity recognition. In CIKM, pages 296–305. ACM.
  12. Lightner: A lightweight tuning paradigm for low-resource NER via pluggable prompting. In COLING, pages 2374–2387. International Committee on Computational Linguistics.
  13. Decoupling knowledge from memorization: Retrieval-augmented prompt learning. In Advances in Neural Information Processing Systems, volume 35, pages 23908–23922. Curran Associates, Inc.
  14. Relation extraction as open-book examination: Retrieval-enhanced prompt tuning. In SIGIR, pages 2443–2448. ACM.
  15. Hybrid transformer with multi-level fusion for multimodal knowledge graph completion. In SIGIR, pages 904–915. ACM.
  16. Good visual guidance makes A better extractor: Hierarchical visual prefix for multimodal entity and relation extraction. In NAACL (Findings).
  17. Knowprompt: Knowledge-aware prompt-tuning with synergistic optimization for relation extraction. In WWW, pages 2778–2788. ACM.
  18. Vistruct: Visual structural knowledge extraction via curriculum guided code-vision representation. In EMNLP. Association for Computational Linguistics.
  19. Prompt-based metric learning for few-shot NER. In ACL (Findings), pages 7199–7212. Association for Computational Linguistics.
  20. Automatically labeled data generation for large scale event extraction. In ACL (1), pages 409–419. Association for Computational Linguistics.
  21. Event extraction via dynamic multi-pooling convolutional neural networks. In ACL (1), pages 167–176. The Association for Computer Linguistics.
  22. Relationprompt: Leveraging prompts to generate synthetic data for zero-shot relation triplet extraction. In ACL (Findings), pages 45–57. Association for Computational Linguistics.
  23. Jason P. C. Chiu and Eric Nichols. 2016. Named entity recognition with bidirectional lstm-cnns. Trans. Assoc. Comput. Linguistics, 4:357–370.
  24. Scaling instruction-finetuned language models. CoRR, abs/2210.11416.
  25. Few-shot event detection with prototypical amortized conditional random field. In ACL/IJCNLP (Findings), volume ACL/IJCNLP 2021 of Findings of ACL, pages 28–40. Association for Computational Linguistics.
  26. James R. Cowie and Wendy G. Lehnert. 1996. Information extraction. Commun. ACM, 39(1):80–91.
  27. Template-based named entity recognition using BART. In ACL/IJCNLP (Findings), volume ACL/IJCNLP 2021 of Findings of ACL, pages 1835–1845. Association for Computational Linguistics.
  28. Unified low-resource sequence labeling by sample-aware dynamic sparse finetuning. In EMNLP. Association for Computational Linguistics.
  29. Speech: Structured prediction with energy-based event-centric hyperspheres. In ACL (1), pages 351–363. Association for Computational Linguistics.
  30. Construction and applications of billion-scale pre-trained multimodal business knowledge graph. In ICDE, pages 2988–3002. IEEE.
  31. Low-resource extraction with knowledge-aware pairwise prototype learning. Knowl. Based Syst., 235:107584.
  32. Meta-learning with dynamic-memory-based prototypical network for few-shot event detection. In WSDM, pages 151–159. ACM.
  33. Ontoed: Low-resource event detection with ontology embedding. In ACL/IJCNLP (1), pages 2828–2839. Association for Computational Linguistics.
  34. When low resource NLP meets unsupervised language model: Meta-pretraining then meta-learning for few-shot text classification (student abstract). In AAAI, pages 13773–13774. AAAI Press.
  35. Knowledge-driven stock trend prediction and explanation via temporal convolutional network. In WWW (Companion Volume), pages 678–685. ACM.
  36. BERT: pre-training of deep bidirectional transformers for language understanding. In NAACL-HLT (1), pages 4171–4186. Association for Computational Linguistics.
  37. Prototypical representation learning for relation extraction. In ICLR. OpenReview.net.
  38. Few-nerd: A few-shot named entity recognition dataset. In ACL/IJCNLP (1), pages 3198–3213. Association for Computational Linguistics.
  39. The automatic content extraction (ACE) program - tasks, data, and evaluation. In LREC. European Language Resources Association.
  40. Meta-information guided meta-learning for few-shot relation classification. In COLING, pages 1594–1605. International Committee on Computational Linguistics.
  41. A multi-task semantic decomposition framework with task-specific pre-training for few-shot NER. In CIKM, pages 430–440. ACM.
  42. Mapre: An effective semantic mapping approach for low-resource relation extraction. In EMNLP (1), pages 2694–2704. Association for Computational Linguistics.
  43. Xinya Du and Claire Cardie. 2020. Event extraction by answering (almost) natural questions. In EMNLP (1), pages 671–683. Association for Computational Linguistics.
  44. Xinya Du and Heng Ji. 2022. Retrieval-augmented generative question answering for event argument extraction. In EMNLP, pages 4649–4666. Association for Computational Linguistics.
  45. GLM: general language model pretraining with autoregressive blank infilling. In ACL (1), pages 320–335. Association for Computational Linguistics.
  46. Multi-sentence argument linking. In ACL, pages 8057–8077. Association for Computational Linguistics.
  47. MANNER: A variational memory-augmented model for cross domain few-shot named entity recognition. In ACL (1), pages 4261–4276. Association for Computational Linguistics.
  48. Lasuie: Unifying information extraction with latent adaptive structure-aware generative language model. In Advances in Neural Information Processing Systems, volume 35, pages 15460–15475. Curran Associates, Inc.
  49. Yuefan Fei and Xiaolong Xu. 2023. GFMRC: A machine reading comprehension model for named entity recognition. Pattern Recognit. Lett., 172:97–105.
  50. Improving low resource named entity recognition using cross-lingual knowledge transfer. In IJCAI, pages 4071–4077. ijcai.org.
  51. Language model priming for cross-lingual event extraction. In AAAI, pages 10627–10635. AAAI Press.
  52. Model-agnostic meta-learning for fast adaptation of deep networks. In ICML, volume 70 of Proceedings of Machine Learning Research, pages 1126–1135. PMLR.
  53. Graphrel: Modeling text as relational graphs for joint entity and relation extraction. In ACL (1), pages 1409–1418. Association for Computational Linguistics.
  54. Mask-then-fill: A flexible and effective data augmentation framework for event extraction. In EMNLP (Findings), pages 4537–4544.
  55. Exploring the feasibility of chatgpt for event extraction. CoRR, abs/2303.03836.
  56. Hybrid attention-based prototypical networks for noisy few-shot relation classification. In AAAI, pages 6407–6414. AAAI Press.
  57. Fewrel 2.0: Towards more challenging few-shot relation classification. In EMNLP/IJCNLP (1), pages 6249–6254. Association for Computational Linguistics.
  58. ACLM: A selective-denoising based generative data augmentation approach for low-resource complex NER. In ACL (1), pages 104–125. Association for Computational Linguistics.
  59. Retrieval-augmented code generation for universal information extraction. CoRR, abs/2311.02962.
  60. Linguistic representations for fewer-shot relation extraction across domains. In ACL (1), pages 7502–7514. Association for Computational Linguistics.
  61. Hybrid knowledge transfer for improved cross-lingual event detection via hierarchical sample selection. In ACL (1), pages 5414–5427. Association for Computational Linguistics.
  62. Is information extraction solved by chatgpt? an analysis of performance, evaluation criteria, robustness and errors. CoRR, abs/2305.14450.
  63. PTR: prompt tuning with rules for text classification. AI Open, 3:182–192.
  64. Fewrel: A large-scale supervised few-shot relation classification dataset with state-of-the-art evaluation. In EMNLP, pages 4803–4809. Association for Computational Linguistics.
  65. A survey on recent approaches for natural language processing in low-resource scenarios. In NAACL-HLT, pages 2545–2568. Association for Computational Linguistics.
  66. Meta-learning in neural networks: A survey. IEEE Trans. Pattern Anal. Mach. Intell., 44(9):5149–5169.
  67. DEGREE: A data-efficient generation-based event extraction model. In NAACL-HLT, pages 1890–1908. Association for Computational Linguistics.
  68. Ampere: Amr-aware prefix for generation-based event argument extraction model. In ACL (1), pages 10976–10993. Association for Computational Linguistics.
  69. Lora: Low-rank adaptation of large language models. In ICLR. OpenReview.net.
  70. Entity-to-text based data augmentation for various named entity recognition tasks. In ACL (Findings), pages 9072–9087. Association for Computational Linguistics.
  71. Gda: Generative data augmentation techniques for relation extraction tasks. In ACL (Findings), pages 10221–10234. Association for Computational Linguistics.
  72. Gradient imitation reinforcement learning for low resource relation extraction. In EMNLP (1), pages 2737–2746. Association for Computational Linguistics.
  73. Event extraction with dynamic prefix tuning and relevance retrieval. IEEE Trans. Knowl. Data Eng., 35(10):9946–9958.
  74. Zero-shot transfer learning for event extraction. In ACL (1), pages 2160–2170. Association for Computational Linguistics.
  75. COPNER: contrastive learning with prompt guiding for few-shot named entity recognition. In COLING, pages 2515–2527. International Committee on Computational Linguistics.
  76. PRAM: an end-to-end prototype-based representation alignment model for zero-resource cross-lingual named entity recognition. In ACL (Findings), pages 3220–3233. Association for Computational Linguistics.
  77. A survey on knowledge graphs: Representation, acquisition, and applications. IEEE Trans. Neural Networks Learn. Syst., 33(2):494–514.
  78. Instruct and extract: Instruction tuning for on-demand information extraction. In EMNLP. Association for Computational Linguistics.
  79. Graph learning regularization and transfer learning for few-shot event detection. In SIGIR, pages 2172–2176. ACM.
  80. BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In ACL, pages 7871–7880. Association for Computational Linguistics.
  81. Evaluating chatgpt’s information extraction capabilities: An assessment of performance, explainability, calibration, and faithfulness. CoRR, abs/2304.11633.
  82. Revisiting large language models as zero-shot relation extractors. In EMNLP (Findings). Association for Computational Linguistics.
  83. Few-shot named entity recognition via meta-learning. IEEE Trans. Knowl. Data Eng., 34(9):4245–4256.
  84. Few-shot relation extraction with dual graph neural network interaction. IEEE Transactions on Neural Networks and Learning Systems, pages 1–13.
  85. Metaner: Named entity recognition with meta-learning. In WWW, pages 429–440.
  86. A survey on deep learning for named entity recognition. IEEE Trans. Knowl. Data Eng., 34(1):50–70.
  87. Logic-guided semantic representation learning for zero-shot relation classification. In COLING, pages 2967–2978. International Committee on Computational Linguistics.
  88. Cross-media structured common space for multimedia event extraction. In ACL, pages 2557–2568. Association for Computational Linguistics.
  89. Connecting the dots: Event graph schema induction with path language modeling. In EMNLP (1), pages 684–695. Association for Computational Linguistics.
  90. Codeie: Large code generation models are better few-shot information extractors. In ACL (1), pages 15339–15353. Association for Computational Linguistics.
  91. A survey on deep learning event extraction: Approaches and applications. IEEE Trans. Neural Networks Learn. Syst., pages 1–21.
  92. Document-level event argument extraction by conditional generation. In NAACL-HLT, pages 894–908. Association for Computational Linguistics.
  93. A unified MRC framework for named entity recognition. In ACL, pages 5849–5859. Association for Computational Linguistics.
  94. Entity-relation extraction as multi-turn question answering. In ACL (1), pages 1340–1350. Association for Computational Linguistics.
  95. BOND: bert-assisted open-domain named entity recognition with distant supervision. In KDD, pages 1054–1064. ACM.
  96. Neural relation extraction with multi-lingual attention. In ACL (1), pages 34–43. Association for Computational Linguistics.
  97. Global constraints with prompting for zero-shot event argument classification. In EACL (Findings), pages 2482–2493. Association for Computational Linguistics.
  98. Improving open information extraction with large language models: A study on demonstration uncertainty. CoRR, abs/2309.03433.
  99. Rexuie: A recursive method with explicit schema instructor for universal information extraction. In EMNLP (Findings). Association for Computational Linguistics.
  100. Pre-training to match for unified low-shot relation extraction. In ACL (1), pages 5785–5795. Association for Computational Linguistics.
  101. Event extraction as machine reading comprehension. In EMNLP (1), pages 1641–1651. Association for Computational Linguistics.
  102. Event detection via gated multilingual attention mechanism. In AAAI, pages 4865–4872. AAAI Press.
  103. Low-resource NER by data augmentation with prompting. In IJCAI, pages 4252–4258. ijcai.org.
  104. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Comput. Surv., 55(9).
  105. Leveraging framenet to improve automatic event detection. In ACL (1). The Association for Computer Linguistics.
  106. Enhancing document-level event argument extraction with contextual clues and role relevance. In ACL (Findings), pages 12908–12922. Association for Computational Linguistics.
  107. Dynamic prefix-tuning for generative template-based event extraction. In ACL (1), pages 5216–5228. Association for Computational Linguistics.
  108. Universal information extraction as unified semantic matching. In AAAI. AAAI Press.
  109. Visual attention model for name tagging in multimodal social media. In ACL (1), pages 1990–1999. Association for Computational Linguistics.
  110. Event extraction as question generation and answering. In ACL (2), pages 1666–1688. Association for Computational Linguistics.
  111. Weiyi Lu and Thien Huu Nguyen. 2018. Similar but not the same - word sense disambiguation improves event detection via neural representation matching. In EMNLP, pages 4822–4828. Association for Computational Linguistics.
  112. Distilling discrimination and generalization knowledge for event detection via delta-representation learning. In ACL (1), pages 4366–4376. Association for Computational Linguistics.
  113. Text2event: Controllable sequence-to-structure generation for end-to-end event extraction. In ACL/IJCNLP (1), pages 2795–2806. Association for Computational Linguistics.
  114. Unified structure generation for universal information extraction. In ACL (1), pages 5755–5772. Association for Computational Linguistics.
  115. DSP: discriminative soft prompts for zero-shot entity and relation extraction. In ACL (Findings), pages 5491–5505. Association for Computational Linguistics.
  116. Label semantics for few shot named entity recognition. In ACL (Findings), pages 1956–1971. Association for Computational Linguistics.
  117. Template-free prompt tuning for few-shot NER. In NAACL-HLT, pages 5721–5732. Association for Computational Linguistics.
  118. Decomposed meta-learning for few-shot named entity recognition. In ACL (Findings), pages 1584–1596. Association for Computational Linguistics.
  119. Interaction information guided prototype representation rectification for few-shot relation extraction. Electronics, 12(13).
  120. Large language model is not a good few-shot information extractor, but a good reranker for hard samples! In EMNLP (Findings). Association for Computational Linguistics.
  121. Prompt for extraction? PAIE: prompting argument interaction for event argument extraction. In ACL (1), pages 6759–6774. Association for Computational Linguistics.
  122. Few-shot event detection: An empirical study and a unified view. In ACL (1), pages 11211–11236. Association for Computational Linguistics.
  123. Jamie P. McCusker. 2023. LOKE: linked open knowledge extraction for automated knowledge graph construction. CoRR, abs/2311.09366.
  124. Distant supervision for relation extraction without labeled data. In ACL/IJCNLP, pages 1003–1011. Association for Computational Linguistics.
  125. Saeed Najafi and Alona Fyshe. 2023. Weakly-supervised questions for zero-shot relation extraction. In EACL, pages 3067–3079. Association for Computational Linguistics.
  126. Retrieving relevant context to align representations for cross-lingual event detection. In ACL (Findings), pages 2157–2170. Association for Computational Linguistics.
  127. Contextualized soft prompts for extraction of event arguments. In ACL (Findings), pages 4352–4361. Association for Computational Linguistics.
  128. DOZEN: cross-domain zero shot named entity recognition with knowledge graph. In SIGIR, pages 1642–1646. ACM.
  129. Knowledge-aware named entity recognition with alleviating heterogeneity. In AAAI, pages 13595–13603. AAAI Press.
  130. Abiola Obamuyide and Andreas Vlachos. 2019. Model-agnostic meta-learning for relation classification with limited supervision. In ACL (1), pages 5873–5879. Association for Computational Linguistics.
  131. Description-based zero-shot fine-grained entity typing. In NAACL-HLT (1), pages 807–814. Association for Computational Linguistics.
  132. OpenAI. 2023. GPT-4 technical report.
  133. In-context few-shot relation extraction via pre-trained language models. arXiv preprint arXiv:2310.11085.
  134. Sinno Jialin Pan and Qiang Yang. 2010. A survey on transfer learning. IEEE Trans. Knowl. Data Eng., 22(10):1345–1359.
  135. Guideline learning for in-context information extraction. In EMNLP. Association for Computational Linguistics.
  136. Structured prediction as translation between augmented natural languages. In ICLR. OpenReview.net.
  137. Bayesian meta-learning for the few-shot setting via deep kernels. In NeurIPS.
  138. Segmix: A simple structure-aware data augmentation method. CoRR, abs/2311.09505.
  139. Robust distant supervision relation extraction via deep reinforcement learning. In ACL (1), pages 2137–2147. Association for Computational Linguistics.
  140. Few-shot relation extraction via bayesian meta-learning on relation graphs. In ICML, volume 119 of Proceedings of Machine Learning Research, pages 7867–7876. PMLR.
  141. Language models are unsupervised multitask learners.
  142. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res., 21:140:1–140:67.
  143. Modeling relations and their mentions without labeled text. In ECML/PKDD (3), volume 6323 of Lecture Notes in Computer Science, pages 148–163. Springer.
  144. Gollie: Annotation guidelines improve zero-shot information-extraction. CoRR, abs/2310.03668.
  145. Erik F. Tjong Kim Sang and Fien De Meulder. 2003. Introduction to the conll-2003 shared task: Language-independent named entity recognition. In CoNLL, pages 142–147. ACL.
  146. Meta-learning with memory-augmented neural networks. In ICML, volume 48 of JMLR Workshop and Conference Proceedings, pages 1842–1850. JMLR.org.
  147. Learning named entity tagger using domain-specific dictionary. In EMNLP, pages 2054–2064. Association for Computational Linguistics.
  148. Adaptive knowledge-enhanced bayesian meta-learning for few-shot event detection. In ACL/IJCNLP (Findings), volume ACL/IJCNLP 2021 of Findings of ACL, pages 2417–2429. Association for Computational Linguistics.
  149. Promptner: Prompt locating and typing for named entity recognition. In ACL (1), pages 12492–12507. Association for Computational Linguistics.
  150. Prototypical networks for few-shot learning. In NIPS, pages 4077–4087.
  151. Matching the blanks: Distributional similarity for relation learning. In ACL (1), pages 2895–2905. Association for Computational Linguistics.
  152. Gpt struct me: Probing gpt models on narrative entity extraction. CoRR, abs/2311.08921.
  153. Nasrin Taghizadeh and Heshaam Faili. 2022. Cross-lingual transfer learning for relation extraction using universal dependencies. Comput. Speech Lang., 71:101265.
  154. Learning structured prediction models: a large margin approach. In ICML, volume 119 of ACM International Conference Proceeding Series, pages 896–903. ACM.
  155. Image enhanced event detection in news articles. In AAAI, pages 9040–9047. AAAI Press.
  156. Improving event detection via open-domain trigger knowledge. In ACL, pages 5887–5897. Association for Computational Linguistics.
  157. Llama: Open and efficient foundation language models. CoRR, abs/2302.13971.
  158. Generating labeled data for relation extraction: A meta learning approach with joint GPT-2 training. In ACL (Findings), pages 11466–11478. Association for Computational Linguistics.
Citations (5)

Summary

We haven't generated a summary for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Lightbulb On Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube