Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Awakening Latent Grounding from Pretrained Language Models for Semantic Parsing (2109.10540v1)

Published 22 Sep 2021 in cs.CL and cs.AI

Abstract: Recent years pretrained LLMs (PLMs) hit a success on several downstream tasks, showing their power on modeling language. To better understand and leverage what PLMs have learned, several techniques have emerged to explore syntactic structures entailed by PLMs. However, few efforts have been made to explore grounding capabilities of PLMs, which are also essential. In this paper, we highlight the ability of PLMs to discover which token should be grounded to which concept, if combined with our proposed erasing-then-awakening approach. Empirical studies on four datasets demonstrate that our approach can awaken latent grounding which is understandable to human experts, even if it is not exposed to such labels during training. More importantly, our approach shows great potential to benefit downstream semantic parsing models. Taking text-to-SQL as a case study, we successfully couple our approach with two off-the-shelf parsers, obtaining an absolute improvement of up to 9.8%.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Qian Liu (252 papers)
  2. Dejian Yang (11 papers)
  3. Jiahui Zhang (64 papers)
  4. Jiaqi Guo (28 papers)
  5. Bin Zhou (161 papers)
  6. Jian-Guang Lou (69 papers)
Citations (39)