GraPPa: Grammar-Augmented Pre-Training for Table Semantic Parsing (2009.13845v2)
Abstract: We present GraPPa, an effective pre-training approach for table semantic parsing that learns a compositional inductive bias in the joint representations of textual and tabular data. We construct synthetic question-SQL pairs over high-quality tables via a synchronous context-free grammar (SCFG) induced from existing text-to-SQL datasets. We pre-train our model on the synthetic data using a novel text-schema linking objective that predicts the syntactic role of a table field in the SQL for each question-SQL pair. To maintain the model's ability to represent real-world data, we also include masked LLMing (MLM) over several existing table-and-language datasets to regularize the pre-training process. On four popular fully supervised and weakly supervised table semantic parsing benchmarks, GraPPa significantly outperforms RoBERTa-large as the feature representation layers and establishes new state-of-the-art results on all of them.
- Tao Yu (282 papers)
- Chien-Sheng Wu (77 papers)
- Xi Victoria Lin (39 papers)
- Bailin Wang (34 papers)
- Yi Chern Tan (9 papers)
- Xinyi Yang (33 papers)
- Dragomir Radev (98 papers)
- Richard Socher (115 papers)
- Caiming Xiong (337 papers)