Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

milIE: Modular & Iterative Multilingual Open Information Extraction (2110.08144v2)

Published 15 Oct 2021 in cs.CL and cs.AI

Abstract: Open Information Extraction (OpenIE) is the task of extracting (subject, predicate, object) triples from natural language sentences. Current OpenIE systems extract all triple slots independently. In contrast, we explore the hypothesis that it may be beneficial to extract triple slots iteratively: first extract easy slots, followed by the difficult ones by conditioning on the easy slots, and therefore achieve a better overall extraction. Based on this hypothesis, we propose a neural OpenIE system, milIE, that operates in an iterative fashion. Due to the iterative nature, the system is also modular -- it is possible to seamlessly integrate rule based extraction systems with a neural end-to-end system, thereby allowing rule based systems to supply extraction slots which milIE can leverage for extracting the remaining slots. We confirm our hypothesis empirically: milIE outperforms SOTA systems on multiple languages ranging from Chinese to Arabic. Additionally, we are the first to provide an OpenIE test dataset for Arabic and Galician.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Bhushan Kotnis (17 papers)
  2. Kiril Gashteovski (19 papers)
  3. Daniel OƱoro Rubio (2 papers)
  4. Vanesa Rodriguez-Tembras (1 paper)
  5. Ammar Shaker (15 papers)
  6. Makoto Takamoto (26 papers)
  7. Mathias Niepert (85 papers)
  8. Carolin Lawrence (29 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.