Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Linking-Enhanced Pre-Training for Table Semantic Parsing (2111.09486v3)

Published 18 Nov 2021 in cs.CL

Abstract: Recently pre-training models have significantly improved the performance of various NLP tasks by leveraging large-scale text corpora to improve the contextual representation ability of the neural network. The large pre-training LLM has also been applied in the area of table semantic parsing. However, existing pre-training approaches have not carefully explored explicit interaction relationships between a question and the corresponding database schema, which is a key ingredient for uncovering their semantic and structural correspondence. Furthermore, the question-aware representation learning in the schema grounding context has received less attention in pre-training objective.To alleviate these issues, this paper designs two novel pre-training objectives to impose the desired inductive bias into the learned representations for table pre-training. We further propose a schema-aware curriculum learning approach to mitigate the impact of noise and learn effectively from the pre-training data in an easy-to-hard manner. We evaluate our pre-trained framework by fine-tuning it on two benchmarks, Spider and SQUALL. The results demonstrate the effectiveness of our pre-training objective and curriculum compared to a variety of baselines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Bowen Qin (16 papers)
  2. Lihan Wang (24 papers)
  3. Binyuan Hui (57 papers)
  4. Ruiying Geng (14 papers)
  5. Zheng Cao (49 papers)
  6. Min Yang (239 papers)
  7. Jian Sun (415 papers)
  8. Yongbin Li (128 papers)
Citations (1)