Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Universal Information Extraction with Meta-Pretrained Self-Retrieval (2306.10444v1)

Published 18 Jun 2023 in cs.CL and cs.AI

Abstract: Universal Information Extraction~(Universal IE) aims to solve different extraction tasks in a uniform text-to-structure generation manner. Such a generation procedure tends to struggle when there exist complex information structures to be extracted. Retrieving knowledge from external knowledge bases may help models to overcome this problem but it is impossible to construct a knowledge base suitable for various IE tasks. Inspired by the fact that large amount of knowledge are stored in the pretrained LLMs~(PLM) and can be retrieved explicitly, in this paper, we propose MetaRetriever to retrieve task-specific knowledge from PLMs to enhance universal IE. As different IE tasks need different knowledge, we further propose a Meta-Pretraining Algorithm which allows MetaRetriever to quicktly achieve maximum task-specific retrieval performance when fine-tuning on downstream IE tasks. Experimental results show that MetaRetriever achieves the new state-of-the-art on 4 IE tasks, 12 datasets under fully-supervised, low-resource and few-shot scenarios.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Xin Cong. Bowen Yu (1 paper)
  2. Mengcheng Fang (1 paper)
  3. Tingwen Liu (45 papers)
  4. Haiyang Yu (109 papers)
  5. Zhongkai Hu (2 papers)
  6. Fei Huang (409 papers)
  7. Yongbin Li (128 papers)
  8. Bin Wang (750 papers)
Citations (4)