Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

From Examples to Rules: Neural Guided Rule Synthesis for Information Extraction (2202.00475v1)

Published 16 Jan 2022 in cs.CL, cs.IR, and cs.LG

Abstract: While deep learning approaches to information extraction have had many successes, they can be difficult to augment or maintain as needs shift. Rule-based methods, on the other hand, can be more easily modified. However, crafting rules requires expertise in linguistics and the domain of interest, making it infeasible for most users. Here we attempt to combine the advantages of these two directions while mitigating their drawbacks. We adapt recent advances from the adjacent field of program synthesis to information extraction, synthesizing rules from provided examples. We use a transformer-based architecture to guide an enumerative search, and show that this reduces the number of steps that need to be explored before a rule is found. Further, we show that without training the synthesis algorithm on the specific domain, our synthesized rules achieve state-of-the-art performance on the 1-shot scenario of a task that focuses on few-shot learning for relation classification, and competitive performance in the 5-shot scenario.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Robert Vacareanu (12 papers)
  2. Marco A. Valenzuela-Escarcega (9 papers)
  3. George C. G. Barbosa (1 paper)
  4. Rebecca Sharp (5 papers)
  5. Mihai Surdeanu (53 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.