Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Large Language Models for Generative Information Extraction: A Survey (2312.17617v3)

Published 29 Dec 2023 in cs.CL
Large Language Models for Generative Information Extraction: A Survey

Abstract: Information extraction (IE) aims to extract structural knowledge from plain natural language texts. Recently, generative LLMs have demonstrated remarkable capabilities in text understanding and generation. As a result, numerous works have been proposed to integrate LLMs for IE tasks based on a generative paradigm. To conduct a comprehensive systematic review and exploration of LLM efforts for IE tasks, in this study, we survey the most recent advancements in this field. We first present an extensive overview by categorizing these works in terms of various IE subtasks and techniques, and then we empirically analyze the most advanced methods and discover the emerging trend of IE tasks with LLMs. Based on a thorough review conducted, we identify several insights in technique and promising research directions that deserve further exploration in future studies. We maintain a public repository and consistently update related works and resources on GitHub (\href{https://github.com/quqxui/Awesome-LLM4IE-Papers}{LLM4IE repository})

Introduction to Generative IE

Generative Information Extraction (IE) is the process of obtaining structured knowledge, like entities, relations, and events, from unstructured textual data. It involves converting text into a structured form that is useful for various downstream applications such as knowledge graph construction, question answering, and knowledge reasoning. LLMs, with their advanced capabilities in text understanding and generation, have paved the way for new methodologies in IE, focusing on generating structured information rather than just extracting it.

Recent Developments in LLMs for IE

LLMs have been integrated into IE tasks with impressive outcomes, and a variety of learning paradigms have been explored to improve their effectiveness. These include supervised fine-tuning, few-shot learning, and zero-shot learning. Each of these paradigms has its own set of strategies tailored to enhance LLMs' performance in IE. For instance, supervised fine-tuning extends a model's capabilities by leveraging existing datasets, while few-shot learning makes use of a small amount of labeled data to train the models effectively. Zero-shot learning employs models that can generalize to new tasks without any labeled examples. Another interesting development is the use of prompts, transforming the extraction process into a query-answering format that guides LLMs in producing the desired output.

The Future of LLMs in IE

This survey has identified several promising directions for future research, emphasizing the potential for the creation of universal frameworks that can handle various IE tasks and domains. Despite advancements, current methods sometimes struggle with long context inputs and structuring outputs in alignment with their training data. Improving the robustness and versatility of these methods is key to unlocking broader applications. Additionally, future research might explore more innovative prompt design strategies that enable LLMs to comprehend tasks more effectively and produce reliable output. Lastly, OpenIE poses unique challenges, suggesting that further investigation is needed to better leverage the knowledge and reasoning abilities of LLMs.

Conclusion

This survey contributes to the understanding of how LLMs are shaping the field of generative IE. By categorizing recent studies according to their learning paradigms and IE tasks, it offers insights into how these models are being fine-tuned, how few-shot or zero-shot learning is applied, and how data augmentation impacts performance. The research suggests that LLMs have much potential in IE, but also underscores the need for ongoing exploration to address existing constraints and unlock the full capabilities of these powerful models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (150)
  1. Large language models are few-shot clinical information extractors. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, pages 1998–2022.
  2. TACRED revisited: A thorough evaluation of the TACRED relation extraction task. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, pages 1558–1569.
  3. Learning to rank context for named entity recognition using a synthetic dataset. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, pages 10372–10382.
  4. Dhananjay Ashok and Zachary C Lipton. 2023. Promptner: Prompting for named entity recognition. arXiv preprint arXiv:2305.15444.
  5. Abductive commonsense reasoning. In 8th International Conference on Learning Representations, ICLR 2020.
  6. Codekgc: Code language model for generative knowledge graph construction. arXiv preprint arXiv:2304.09048.
  7. Inspire the large language model by external knowledge on biomedical named entity recognition. arXiv preprint arXiv:2309.12278.
  8. Language models are few-shot learners. 33:1877–1901.
  9. Pere-Lluís Huguet Cabot and Roberto Navigli. 2021. Rebel: Relation extraction by end-to-end language generation. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 2370–2381.
  10. In-context learning for few-shot multimodal named entity recognition. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 2969–2979.
  11. Erica Cai and Brendan O’Connor. 2023. A monte carlo language model pipeline for zero-shot sociopolitical event extraction. arXiv preprint arXiv:2305.15051.
  12. Feng Chen and Yujian Feng. 2023. Chain-of-thought prompt distillation for multimodal named entity and multimodal relation extraction. arXiv preprint arXiv:2306.14122.
  13. Crossroads, buildings and neighborhoods: A dataset for fine-grained location recognition. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2022, pages 3329–3339.
  14. Heproto: A hierarchical enhancing protonet based on multi-task learning for few-shot named entity recognition. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, CIKM 2023, pages 296–305.
  15. One model for all domains: Collaborative domain-prefix tuning for cross-domain NER. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI 2023, pages 5030–5038.
  16. Polyie: A dataset of information extraction from polymer material scientific literature. arXiv preprint arXiv:2311.07715.
  17. Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality. See https://vicuna. lmsys. org (accessed 14 April 2023).
  18. Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416.
  19. Template-based named entity recognition using bart. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 1835–1845.
  20. Broad twitter corpus: A diverse named entity recognition resource. In COLING 2016, 26th International Conference on Computational Linguistics, Proceedings of the Conference: Technical Papers, pages 1169–1179.
  21. Results of the WNUT2017 shared task on novel and emerging entity recognition. In Proceedings of the 3rd Workshop on Noisy User-generated Text, NUT@EMNLP 2017, pages 140–147.
  22. Few-nerd: A few-shot named entity recognition dataset. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), pages 3198–3213.
  23. The automatic content extraction (ACE) program - tasks, data, and evaluation. In Proceedings of the Fourth International Conference on Language Resources and Evaluation, LREC 2004. European Language Resources Association.
  24. Ncbi disease corpus: a resource for disease name recognition and concept normalization. Journal of biomedical informatics, 47:1–10.
  25. Glm: General language model pretraining with autoregressive blank infilling. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, pages 320–335.
  26. Structured information extraction from complex scientific text with fine-tuned large language models. arXiv preprint arXiv:2212.05238.
  27. Multi-sentence argument linking. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, pages 8057–8077.
  28. Calibrated seq2seq models for efficient and generalizable ultra-fine entity typing. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 15550–15560.
  29. Collaborative policy learning for open knowledge graph reasoning. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, pages 2672–2681.
  30. Giellm: Japanese general information extraction large language model utilizing mutual reinforcement effect. arXiv preprint arXiv:2311.06838.
  31. Exploring the feasibility of chatgpt for event extraction. arXiv preprint arXiv:2303.03836.
  32. Yes but.. can chatgpt identify entities in historical documents? In ACM/IEEE Joint Conference on Digital Libraries, JCDL 2023, pages 184–189.
  33. Findvehicle and vehiclefinder: A ner dataset for natural language-based vehicle retrieval and a keyword-based cross-modal vehicle retrieval system. arXiv preprint arXiv:2304.10893.
  34. Instructie: A chinese instruction-based information extraction dataset. arXiv preprint arXiv:2305.11527.
  35. Retrieval-augmented code generation for universal information extraction. arXiv preprint arXiv:2311.02962.
  36. Development of a benchmark corpus to support the automatic extraction of drug-related adverse effects from medical case reports. J. Biomed. Informatics, 45(5):885–892.
  37. Thinking about gpt-3 in-context learning for biomedical ie? think again. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 4497–4512.
  38. Is information extraction solved by chatgpt? an analysis of performance, evaluation criteria, robustness and errors. arXiv preprint arXiv:2305.14450.
  39. Semeval-2010 task 8: Multi-way classification of semantic relations between pairs of nominals. In Proceedings of the 5th International Workshop on Semantic Evaluation, SemEval@ACL 2010, pages 33–38.
  40. Degree: A data-efficient generation-based event extraction model. arXiv preprint arXiv:2108.12724.
  41. Entity-to-text based data augmentation for various named entity recognition tasks. In Findings of the Association for Computational Linguistics: ACL 2023, pages 9072–9087.
  42. Zero-shot clinical entity recognition using chatgpt. arXiv preprint arXiv:2303.16416.
  43. Few-shot named entity recognition: A comprehensive study. arXiv preprint arXiv:2012.14978.
  44. Multilingual generative language models for zero-shot cross-lingual event argument extraction. arXiv preprint arXiv:2203.08308.
  45. Document-level entity-based extraction as template generation. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, pages 5257–5269.
  46. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288.
  47. Annotating the tweebank corpus on named entity recognition and building NLP models for social media analysis. In Proceedings of the Thirteenth Language Resources and Evaluation Conference, LREC 2022, pages 7199–7208.
  48. Genie: Generative information extraction. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2022, pages 4626–4643.
  49. Exploiting asymmetry for synthetic training data generation: Synthie and the case of information extraction. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, pages 1555–1574.
  50. Cadec: A corpus of adverse drug event annotations. J. Biomed. Informatics, 55:73–81.
  51. Neretrieve: Dataset for next generation named entity recognition and retrieval. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 3340–3354.
  52. GENIA corpus - a semantically annotated corpus for bio-textmining. In Proceedings of the Eleventh International Conference on Intelligent Systems for Molecular Biology, pages 180–182.
  53. Overview of genia event task in bionlp shared task 2011. In Proceedings of BioNLP Shared Task 2011 Workshop, pages 7–15.
  54. The genia event extraction shared task, 2013 edition - overview. In Proceedings of the BioNLP Shared Task 2013 Workshop, pages 8–15.
  55. Large language models are zero-shot reasoners. Advances in neural information processing systems, 35:22199–22213.
  56. Aman Kumar and Binil Starly. 2022. "fabner": information extraction from manufacturing process science domain literature using named entity recognition. J. Intell. Manuf., 33(8):2393–2407.
  57. A zero-shot and few-shot study of instruction-finetuned large language models applied to clinical and biomedical tasks. arXiv preprint arXiv:2307.12114.
  58. Evaluating chatgpt’s information extraction capabilities: An assessment of performance, explainability, calibration, and faithfulness. arXiv preprint arXiv:2304.11633.
  59. Revisiting large language models as zero-shot relation extractors. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 6877–6892.
  60. Set learning for generative information extraction. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, pages 13043–13052.
  61. Biocreative v cdr task corpus: a resource for chemical disease relation extraction. Database J. Biol. Databases Curation, 2016.
  62. Prompting chatgpt in MNER: enhanced multimodal named entity recognition with auxiliary refined knowledge. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 2787–2802.
  63. Semi-automatic data enhancement for document-level relation extraction with distant supervision from large language models. arXiv preprint arXiv:2311.07314.
  64. Mingchen Li and Rui Zhang. 2023. How far is language model from 100% few-shot named entity recognition in medical domain. arXiv preprint arXiv:2307.00186.
  65. Codeie: Large code generation models are better few-shot information extractors. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, pages 15339–15353. Association for Computational Linguistics.
  66. Document-level event argument extraction by conditional generation. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021, pages 894–908.
  67. Rexuie: A recursive method with explicit schema instructor for universal information extraction. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 15342–15359.
  68. Asgard: A portable architecture for multilingual dialogue systems. In IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2013, pages 8386–8390. IEEE.
  69. Dynamic prefix-tuning for generative template-based event extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, pages 5216–5228.
  70. Crossner: Evaluating cross-domain named entity recognition. In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, volume 35, pages 13452–13460.
  71. Universal information extraction as unified semantic matching. In Thirty-Seventh AAAI Conference on Artificial Intelligence, AAAI 2023, Thirty-Fifth Conference on Innovative Applications of Artificial Intelligence, IAAI 2023, Thirteenth Symposium on Educational Advances in Artificial Intelligence, EAAI 2023, pages 13318–13326.
  72. Visual attention model for name tagging in multimodal social media. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2018, pages 1990–1999.
  73. Event extraction as question generation and answering. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), ACL 2023, pages 1666–1688.
  74. Text2event: Controllable sequence-to-structure generation for end-to-end event extraction. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), ACL 2021, pages 2795–2806.
  75. Unified structure generation for universal information extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, pages 5755–5772.
  76. Multi-task identification of entities, relations, and coreference for scientific knowledge graph construction. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018, pages 3219–3232.
  77. Lei Luo and Yajing Xu. 2023. Context-aware prompt for generation-based event argument extraction with diffusion models. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, CIKM 2023, pages 1717–1725.
  78. DICE: data-efficient clinical event extraction with generative models. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, pages 15898–15917. Association for Computational Linguistics.
  79. Star: Boosting low-resource event extraction by structure-to-text data generation with large language models. arXiv preprint arXiv:2305.15090.
  80. Large language model is not a good few-shot information extractor, but a good reranker for hard samples! In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 10572–10601.
  81. Prompt for extraction? paie: Prompting argument interaction for event argument extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, pages 6759–6774.
  82. Large language models as instructors: A study on multilingual clinical entity extraction. In The 22nd Workshop on Biomedical Natural Language Processing and BioNLP Shared Tasks, pages 178–190.
  83. Task 2: Share/clef ehealth evaluation lab 2014. In Working Notes for CLEF 2014 Conference, volume 1180 of CEUR Workshop Proceedings, pages 31–42.
  84. Named entity recognition and relation extraction: State-of-the-art. ACM Computing Surveys (CSUR), 54(1):1–39.
  85. OpenAI. 2023a. Gpt-4 technical report. ArXiv, abs/2303.08774.
  86. OpenAI. 2023b. Introduce chatgpt. OpenAI blog.
  87. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
  88. Guideline learning for in-context information extraction. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, pages 15372–15389.
  89. Structured prediction as translation between augmented natural languages. In 9th International Conference on Learning Representations, ICLR 2021.
  90. Task 1: Share/clef ehealth evaluation lab 2013. In Working Notes for CLEF 2013 Conference, volume 1179 of CEUR Workshop Proceedings.
  91. Towards robust linguistic analysis using ontonotes. In Proceedings of the Seventeenth Conference on Computational Natural Language Learning, CoNLL 2013, pages 143–152.
  92. Preserving knowledge invariance: Rethinking robustness evaluation of open information extraction. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, pages 5876–5890.
  93. Reasoning with language model prompting: A survey. arXiv preprint arXiv:2212.09597.
  94. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9.
  95. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res., 21:140:1–140:67.
  96. Modeling relations and their mentions without labeled text. In Machine Learning and Knowledge Discovery in Databases, European Conference, ECML PKDD 2010, volume 6323 of Lecture Notes in Computer Science, pages 148–163.
  97. Shruti Rijhwani and Daniel Preotiuc-Pietro. 2020. Temporally-informed analysis of named entity recognition. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, pages 7605–7617.
  98. Dan Roth and Wen-tau Yih. 2004. A linear programming formulation for global inference in natural language tasks. In Proceedings of the Eighth Conference on Computational Natural Language Learning, CoNLL 2004, Held in cooperation with HLT-NAACL 2004, pages 1–8.
  99. Code llama: Open foundation models for code. arXiv preprint arXiv:2308.12950.
  100. Gollie: Annotation guidelines improve zero-shot information-extraction. arXiv preprint arXiv:2310.03668.
  101. Erik Tjong Kim Sang and Fien De Meulder. 2003. Introduction to the conll-2003 shared task: Language-independent named entity recognition. In Proceedings of the Seventh Conference on Natural Language Learning, CoNLL 2003, Held in cooperation with HLT-NAACL 2003, pages 142–147.
  102. CASIE: extracting cybersecurity event information from text. In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, pages 8749–8757.
  103. Prompt-ner: Zero-shot named entity recognition in astronomy literature via large language models. arXiv preprint arXiv:2310.17892.
  104. Information extraction supported question answering. In TREC.
  105. Re-tacred: Addressing shortcomings of the TACRED dataset. In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, pages 13843–13850. AAAI Press.
  106. PHEE: A dataset for pharmacovigilance event extraction from text. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, pages 5571–5587.
  107. Does synthetic data generation of llms help clinical text mining? arXiv preprint arXiv:2303.04360.
  108. Stanford alpaca: An instruction-following llama model.
  109. Simone Tedeschi and Roberto Navigli. 2022. Multinerd: A multilingual, multi-genre and fine-grained dataset for named entity recognition (and disambiguation). In Findings of the Association for Computational Linguistics: NAACL 2022, pages 801–812.
  110. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971.
  111. Named entity recognition in twitter: A dataset and analysis on short-term temporal shifts. In Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing, AACL/IJCNLP 2022 - Volume 1: Long Papers, pages 309–319.
  112. Unleash gpt-2 power for event detection. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 6271–6282.
  113. Revisiting relation extraction in the era of large language models. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, pages 15566–15589.
  114. Ace 2005 multilingual training corpus ldc2006t06, 2006. URL https://catalog. ldc. upenn. edu/LDC2006T06.
  115. GPT-RE: in-context learning for relation extraction using large language models. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, pages 3534–3547.
  116. Ben Wang. 2021. Mesh-Transformer-JAX: Model-Parallel Implementation of Transformer Language Model with JAX. https://github.com/kingoflolz/mesh-transformer-jax.
  117. Deepstruct: Pretraining of language models for structure prediction. In Findings of the Association for Computational Linguistics: ACL 2022, pages 803–823.
  118. Improving unsupervised relation extraction by augmenting diverse sentence pairs. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, pages 12136–12147.
  119. Gpt-ner: Named entity recognition via large language models. arXiv preprint arXiv:2304.10428.
  120. Instructuie: Multi-task instruction tuning for unified information extraction. arXiv preprint arXiv:2304.08085.
  121. Code4struct: Code generation for few-shot event structure prediction. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, pages 3640–3663.
  122. Wikidiverse: A multimodal entity linking dataset with diversified contextual topics and entity types. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, pages 4785–4797.
  123. Crossweigh: Training named entity tagger from imperfect annotations. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, pages 5153–5162.
  124. Zero-shot information extraction via chatting with chatgpt. arXiv preprint arXiv:2302.10205.
  125. Llm-powered data augmentation for enhanced crosslingual performance. arXiv preprint arXiv:2305.14288.
  126. Dafs: a domain aware few shot generative model for event detection. Machine Learning, 112(3):1011–1031.
  127. Debiasing generative named entity recognition by calibrating sequence likelihood. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), ACL 2023, pages 1137–1148.
  128. Empirical study of zero-shot ner with chatgpt. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, pages 7935–7956.
  129. Self-improving for zero-shot named entity recognition with large language models. arXiv preprint arXiv:2311.08921.
  130. How to unleash the power of large language models for few-shot relation extraction? In Proceedings of The Fourth Workshop on Simple and Efficient Natural Language Processing, SustaiNLP 2023, pages 190–200. Association for Computational Linguistics.
  131. mt5: A massively multilingual pre-trained text-to-text transformer. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021, pages 483–498.
  132. A unified generative framework for various ner subtasks. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), pages 5808–5822.
  133. Docred: A large-scale document-level relation extraction dataset. In Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, pages 764–777.
  134. Generative knowledge graph construction: A review. arXiv preprint arXiv:2210.12714.
  135. A survey on multimodal large language models. arXiv preprint arXiv:2306.13549.
  136. Zero-shot temporal relation extraction with chatgpt. In The 22nd Workshop on Biomedical Natural Language Processing and BioNLP Shared Tasks, BioNLP@ACL 2023, pages 92–102.
  137. Generative entity typing with curriculum learning. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, pages 3061–3073.
  138. 2iner: Instructive and in-context learning on few-shot named entity recognition. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 3940–3951.
  139. Aligning instruction tasks unlocks large language models as zero-shot relation extractors. In Findings of the Association for Computational Linguistics: ACL 2023, pages 794–812.
  140. Adaptive co-attention network for named entity recognition in tweets. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence, pages 5674–5681.
  141. Llmaaa: Making large language models as active annotators. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 13088–13103.
  142. De-bias for generative extraction in unified ner task. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, pages 808–818.
  143. Position-aware attention and supervised data improve slot filling. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, EMNLP 2017, pages 35–45.
  144. A comprehensive survey on deep learning for relation extraction: Recent advances and new frontiers. arXiv e-prints, pages arXiv–2306.
  145. Mnre: A challenge multimodal dataset for neural relation extraction with visual evidence in social media posts. In 2021 IEEE International Conference on Multimedia and Expo, ICME 2021, pages 1–6. IEEE.
  146. A comprehensive survey on automatic knowledge graph construction. arXiv preprint arXiv:2302.05019.
  147. A survey on neural open information extraction: Current status and future directions. arXiv preprint arXiv:2205.11725.
  148. Universalner: Targeted distillation from large language models for open named entity recognition. arXiv preprint arXiv:2308.03279.
  149. Claret: Pre-training a correlation-aware context-to-event transformer for event-centric generation and classification. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, pages 2559–2575.
  150. Mirror: A universal framework for various information extraction tasks. arXiv preprint arXiv:2311.05419.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Derong Xu (26 papers)
  2. Wei Chen (1288 papers)
  3. Wenjun Peng (6 papers)
  4. Chao Zhang (907 papers)
  5. Tong Xu (113 papers)
  6. Xiangyu Zhao (192 papers)
  7. Xian Wu (139 papers)
  8. Yefeng Zheng (197 papers)
  9. Enhong Chen (242 papers)
  10. Yang Wang (670 papers)
Citations (92)