Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Contextual Representations for Semantic Parsing with Generation-Augmented Pre-Training (2012.10309v1)

Published 18 Dec 2020 in cs.CL

Abstract: Most recently, there has been significant interest in learning contextual representations for various NLP tasks, by leveraging large scale text corpora to train large neural LLMs with self-supervised learning objectives, such as Masked LLM (MLM). However, based on a pilot study, we observe three issues of existing general-purpose LLMs when they are applied to text-to-SQL semantic parsers: fail to detect column mentions in the utterances, fail to infer column mentions from cell values, and fail to compose complex SQL queries. To mitigate these issues, we present a model pre-training framework, Generation-Augmented Pre-training (GAP), that jointly learns representations of natural language utterances and table schemas by leveraging generation models to generate pre-train data. GAP MODEL is trained on 2M utterance-schema pairs and 30K utterance-schema-SQL triples, whose utterances are produced by generative models. Based on experimental results, neural semantic parsers that leverage GAP MODEL as a representation encoder obtain new state-of-the-art results on both SPIDER and CRITERIA-TO-SQL benchmarks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Peng Shi (80 papers)
  2. Patrick Ng (29 papers)
  3. Zhiguo Wang (100 papers)
  4. Henghui Zhu (24 papers)
  5. Alexander Hanbo Li (17 papers)
  6. Jun Wang (991 papers)
  7. Cicero Nogueira dos Santos (31 papers)
  8. Bing Xiang (74 papers)
Citations (110)