ChatGPT for Zero-shot Dialogue State Tracking: A Solution or an Opportunity? (2306.01386v1)
Abstract: Recent research on dialogue state tracking (DST) focuses on methods that allow few- and zero-shot transfer to new domains or schemas. However, performance gains heavily depend on aggressive data augmentation and fine-tuning of ever larger LLM based architectures. In contrast, general purpose LLMs, trained on large amounts of diverse data, hold the promise of solving any kind of task without task-specific training. We present preliminary experimental results on the ChatGPT research preview, showing that ChatGPT achieves state-of-the-art performance in zero-shot DST. Despite our findings, we argue that properties inherent to general purpose models limit their ability to replace specialized systems. We further theorize that the in-context learning capabilities of such models will likely become powerful tools to support the development of dedicated and dynamic dialogue state trackers.
- Michael Heck (23 papers)
- Nurul Lubis (21 papers)
- Benjamin Ruppik (11 papers)
- Renato Vukovic (10 papers)
- Shutong Feng (19 papers)
- Christian Geishauser (19 papers)
- Carel van Niekerk (23 papers)
- Milica Gašić (57 papers)
- Hsien-chin Lin (22 papers)