Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Decoupled Dialogue Modeling and Semantic Parsing for Multi-Turn Text-to-SQL (2106.02282v2)

Published 4 Jun 2021 in cs.CL

Abstract: Recently, Text-to-SQL for multi-turn dialogue has attracted great interest. Here, the user input of the current turn is parsed into the corresponding SQL query of the appropriate database, given all previous dialogue history. Current approaches mostly employ end-to-end models and consequently face two challenges. First, dialogue history modeling and Text-to-SQL parsing are implicitly combined, hence it is hard to carry out interpretable analysis and obtain targeted improvement. Second, SQL annotation of multi-turn dialogue is very expensive, leading to training data sparsity. In this paper, we propose a novel decoupled multi-turn Text-to-SQL framework, where an utterance rewrite model first explicitly solves completion of dialogue context, and then a single-turn Text-to-SQL parser follows. A dual learning approach is also proposed for the utterance rewrite model to address the data sparsity problem. Compared with end-to-end approaches, the proposed decoupled method can achieve excellent performance without any annotated in-domain data. With just a few annotated rewrite cases, the decoupled method outperforms the released state-of-the-art end-to-end models on both SParC and CoSQL datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Zhi Chen (235 papers)
  2. Lu Chen (245 papers)
  3. Hanqi Li (9 papers)
  4. Ruisheng Cao (24 papers)
  5. Da Ma (28 papers)
  6. Mengyue Wu (57 papers)
  7. Kai Yu (202 papers)
Citations (18)

Summary

We haven't generated a summary for this paper yet.