Linking-Enhanced Pre-Training for Table Semantic Parsing (2111.09486v3)
Abstract: Recently pre-training models have significantly improved the performance of various NLP tasks by leveraging large-scale text corpora to improve the contextual representation ability of the neural network. The large pre-training LLM has also been applied in the area of table semantic parsing. However, existing pre-training approaches have not carefully explored explicit interaction relationships between a question and the corresponding database schema, which is a key ingredient for uncovering their semantic and structural correspondence. Furthermore, the question-aware representation learning in the schema grounding context has received less attention in pre-training objective.To alleviate these issues, this paper designs two novel pre-training objectives to impose the desired inductive bias into the learned representations for table pre-training. We further propose a schema-aware curriculum learning approach to mitigate the impact of noise and learn effectively from the pre-training data in an easy-to-hard manner. We evaluate our pre-trained framework by fine-tuning it on two benchmarks, Spider and SQUALL. The results demonstrate the effectiveness of our pre-training objective and curriculum compared to a variety of baselines.
- Bowen Qin (16 papers)
- Lihan Wang (24 papers)
- Binyuan Hui (57 papers)
- Ruiying Geng (14 papers)
- Zheng Cao (49 papers)
- Min Yang (239 papers)
- Jian Sun (415 papers)
- Yongbin Li (128 papers)