Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Mastering the Task of Open Information Extraction with Large Language Models and Consistent Reasoning Environment (2310.10590v1)

Published 16 Oct 2023 in cs.CL

Abstract: Open Information Extraction (OIE) aims to extract objective structured knowledge from natural texts, which has attracted growing attention to build dedicated models with human experience. As the LLMs have exhibited remarkable in-context learning capabilities, a question arises as to whether the task of OIE can be effectively tackled with this paradigm? In this paper, we explore solving the OIE problem by constructing an appropriate reasoning environment for LLMs. Specifically, we first propose a method to effectively estimate the discrepancy of syntactic distribution between a LLM and test samples, which can serve as correlation evidence for preparing positive demonstrations. Upon the evidence, we introduce a simple yet effective mechanism to establish the reasoning environment for LLMs on specific tasks. Without bells and whistles, experimental results on the standard CaRB benchmark demonstrate that our $6$-shot approach outperforms state-of-the-art supervised method, achieving an $55.3$ $F_1$ score. Further experiments on TACRED and ACE05 show that our method can naturally generalize to other information extraction tasks, resulting in improvements of $5.7$ and $6.8$ $F_1$ scores, respectively.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Ji Qi (61 papers)
  2. Kaixuan Ji (11 papers)
  3. Xiaozhi Wang (51 papers)
  4. Jifan Yu (49 papers)
  5. Kaisheng Zeng (17 papers)
  6. Lei Hou (127 papers)
  7. Juanzi Li (144 papers)
  8. Bin Xu (192 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.