Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SynthDST: Synthetic Data is All You Need for Few-Shot Dialog State Tracking (2402.02285v1)

Published 3 Feb 2024 in cs.CL, cs.AI, and cs.LG

Abstract: In-context learning with LLMs has emerged as a promising avenue of research in Dialog State Tracking (DST). However, the best-performing in-context learning methods involve retrieving and adding similar examples to the prompt, requiring access to labeled training data. Procuring such training data for a wide range of domains and applications is time-consuming, expensive, and, at times, infeasible. While zero-shot learning requires no training data, it significantly lags behind the few-shot setup. Thus, `\textit{Can we efficiently generate synthetic data for any dialogue schema to enable few-shot prompting?}' Addressing this question, we propose \method, a data generation framework tailored for DST, utilizing LLMs. Our approach only requires the dialogue schema and a few hand-crafted dialogue templates to synthesize natural, coherent, and free-flowing dialogues with DST annotations. Few-shot learning using data from {\method} results in $4-5%$ improvement in Joint Goal Accuracy over the zero-shot baseline on MultiWOZ 2.1 and 2.4. Remarkably, our few-shot learning approach recovers nearly $98%$ of the performance compared to the few-shot setup using human-annotated training data. Our synthetic data and code can be accessed at https://github.com/apple/ml-synthdst

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Atharva Kulkarni (28 papers)
  2. Bo-Hsiang Tseng (20 papers)
  3. Joel Ruben Antony Moniz (23 papers)
  4. Dhivya Piraviperumal (8 papers)
  5. Hong Yu (114 papers)
  6. Shruti Bhargava (10 papers)
Citations (5)
Youtube Logo Streamline Icon: https://streamlinehq.com