Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Sample-efficient Linguistic Generalizations through Program Synthesis: Experiments with Phonology Problems (2106.06566v1)

Published 11 Jun 2021 in cs.CL

Abstract: Neural models excel at extracting statistical patterns from large amounts of data, but struggle to learn patterns or reason about language from only a few examples. In this paper, we ask: Can we learn explicit rules that generalize well from only a few examples? We explore this question using program synthesis. We develop a synthesis model to learn phonology rules as programs in a domain-specific language. We test the ability of our models to generalize from few training examples using our new dataset of problems from the Linguistics Olympiad, a challenging set of tasks that require strong linguistic reasoning ability. In addition to being highly sample-efficient, our approach generates human-readable programs, and allows control over the generalizability of the learnt programs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Saujas Vaduguru (8 papers)
  2. Aalok Sathe (4 papers)
  3. Monojit Choudhury (66 papers)
  4. Dipti Misra Sharma (15 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.