Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Syntactic and Semantic-driven Learning for Open Information Extraction (2103.03448v1)

Published 5 Mar 2021 in cs.CL

Abstract: One of the biggest bottlenecks in building accurate, high coverage neural open IE systems is the need for large labelled corpora. The diversity of open domain corpora and the variety of natural language expressions further exacerbate this problem. In this paper, we propose a syntactic and semantic-driven learning approach, which can learn neural open IE models without any human-labelled data by leveraging syntactic and semantic knowledge as noisier, higher-level supervisions. Specifically, we first employ syntactic patterns as data labelling functions and pretrain a base model using the generated labels. Then we propose a syntactic and semantic-driven reinforcement learning algorithm, which can effectively generalize the base model to open situations with high accuracy. Experimental results show that our approach significantly outperforms the supervised counterparts, and can even achieve competitive performance to supervised state-of-the-art (SoA) model

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Jialong Tang (17 papers)
  2. Yaojie Lu (61 papers)
  3. Hongyu Lin (94 papers)
  4. Xianpei Han (103 papers)
  5. Le Sun (111 papers)
  6. Xinyan Xiao (41 papers)
  7. Hua Wu (191 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.