Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

In-Context Learning for Few-Shot Dialogue State Tracking (2203.08568v3)

Published 16 Mar 2022 in cs.CL

Abstract: Collecting and annotating task-oriented dialogues is time-consuming and costly; thus, zero and few shot learning could greatly benefit dialogue state tracking (DST). In this work, we propose an in-context learning (ICL) framework for zero-shot and few-shot learning DST, where a large pre-trained LLM (LM) takes a test instance and a few exemplars as input, and directly decodes the dialogue state without any parameter updates. To better leverage a tabular domain description in the LM prompt, we reformulate DST into a text-to-SQL problem. We also propose a novel approach to retrieve annotated dialogues as exemplars. Empirical results on MultiWOZ show that our method IC-DST substantially outperforms previous fine-tuned state-of-the-art models in few-shot settings. In addition, we test IC-DST in zero-shot settings, in which the model only takes a fixed task instruction as input, finding that it outperforms previous zero-shot methods by a large margin.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yushi Hu (23 papers)
  2. Chia-Hsuan Lee (12 papers)
  3. Tianbao Xie (22 papers)
  4. Tao Yu (282 papers)
  5. Noah A. Smith (224 papers)
  6. Mari Ostendorf (57 papers)
Citations (53)
Github Logo Streamline Icon: https://streamlinehq.com