Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

InstructUIE: Multi-task Instruction Tuning for Unified Information Extraction (2304.08085v1)

Published 17 Apr 2023 in cs.CL and cs.AI
InstructUIE: Multi-task Instruction Tuning for Unified Information Extraction

Abstract: LLMs have unlocked strong multi-task capabilities from reading instructive prompts. However, recent studies have shown that existing large models still have difficulty with information extraction tasks. For example, gpt-3.5-turbo achieved an F1 score of 18.22 on the Ontonotes dataset, which is significantly lower than the state-of-the-art performance. In this paper, we propose InstructUIE, a unified information extraction framework based on instruction tuning, which can uniformly model various information extraction tasks and capture the inter-task dependency. To validate the proposed method, we introduce IE INSTRUCTIONS, a benchmark of 32 diverse information extraction datasets in a unified text-to-text format with expert-written instructions. Experimental results demonstrate that our method achieves comparable performance to Bert in supervised settings and significantly outperforms the state-of-the-art and gpt3.5 in zero-shot settings.

InstructUIE: Multi-task Instruction Tuning for Unified Information Extraction

The paper presents InstructUIE, a novel framework that leverages instruction tuning to create a unified information extraction (IE) model. This model aims to address the limitations of LLMs in IE tasks, where models like gpt-3.5-turbo underperform significantly compared to state-of-the-art benchmarks. The proposed InstructUIE framework reformulates IE tasks as a text-to-text problem and integrates auxiliary tasks to enhance its structural understanding.

Methodology

InstructUIE utilizes instruction tuning, a multi-task learning framework, which provides human-readable prompts to guide LLM outputs. The framework involves:

  • Task Schema: Each task instance in InstructUIE is composed of detailed task instructions, options (output constraints), input text, and required textual outputs. Tasks are transformed into sequence-to-sequence (seq2seq) formats.
  • Auxiliary Tasks: Complementary tasks such as entity span extraction and entity typing for NER, and entity pair extraction for RE, improve the model’s understanding of task-specific structures.
  • IE INSTRUCTIONS Benchmark: A comprehensive benchmark of 32 datasets in standardized text-to-text formats for evaluating IE tasks.

Experimental Results

The experimental evaluation of InstructUIE covered supervised and zero-shot settings on numerous datasets across NER, RE, and EE tasks.

Supervised Settings

  • Named Entity Recognition (NER): InstructUIE outperformed Bert on 17 out of 20 datasets, achieving an average F1 score of 85.19%. Particularly notable was the broad twitter dataset, where it surpassed Bert by 25 points.
  • Relation Extraction (RE) and Event Extraction (EE): Results indicated competitive performance, with InstructUIE achieving higher F1 scores than UIE and USM on many datasets.

Zero-shot Settings

  • NER: InstructUIE excelled over USM across most datasets, with improvements ranging between 5.21% to 25.27% in Micro-F1 scores.
  • RE: InstructUIE outperformed ZETT on FewRel and Wiki-ZSL datasets by a margin of over 5% in F1 scores.

Implications and Future Directions

InstructUIE demonstrates the potential for instruction-tuned LLMs in effectively handling various IE tasks with limited task-specific data. The ability of InstructUIE to utilize multi-task learning frameworks opens up new possibilities for improving generalization across unseen tasks, especially in zero-shot contexts. Future developments in AI could explore enhancing instruction sets and adapting instruction tuning for a broader spectrum of natural language processing challenges. This approach also holds promise for reducing the dependency on extensive labeled datasets, expediting the deployment of IE systems in novel domains. Overall, InstructUIE represents a substantial step forward in the pursuit of universal models for information extraction.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (14)
  1. Xiao Wang (507 papers)
  2. Weikang Zhou (10 papers)
  3. Can Zu (5 papers)
  4. Han Xia (6 papers)
  5. Tianze Chen (9 papers)
  6. Yuansen Zhang (6 papers)
  7. Rui Zheng (78 papers)
  8. Junjie Ye (66 papers)
  9. Qi Zhang (784 papers)
  10. Tao Gui (127 papers)
  11. Jihua Kang (2 papers)
  12. Jingsheng Yang (2 papers)
  13. Siyuan Li (140 papers)
  14. Chunsai Du (1 paper)
Citations (119)
Github Logo Streamline Icon: https://streamlinehq.com