Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GALAXY: A Generative Pre-trained Model for Task-Oriented Dialog with Semi-Supervised Learning and Explicit Policy Injection (2111.14592v8)

Published 29 Nov 2021 in cs.CL

Abstract: Pre-trained models have proved to be powerful in enhancing task-oriented dialog systems. However, current pre-training methods mainly focus on enhancing dialog understanding and generation tasks while neglecting the exploitation of dialog policy. In this paper, we propose GALAXY, a novel pre-trained dialog model that explicitly learns dialog policy from limited labeled dialogs and large-scale unlabeled dialog corpora via semi-supervised learning. Specifically, we introduce a dialog act prediction task for policy optimization during pre-training and employ a consistency regularization term to refine the learned representation with the help of unlabeled dialogs. We also implement a gating mechanism to weigh suitable unlabeled dialog samples. Empirical results show that GALAXY substantially improves the performance of task-oriented dialog systems, and achieves new state-of-the-art results on benchmark datasets: In-Car, MultiWOZ2.0 and MultiWOZ2.1, improving their end-to-end combined scores by 2.5, 5.3 and 5.5 points, respectively. We also show that GALAXY has a stronger few-shot ability than existing models under various low-resource settings.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (12)
  1. Wanwei He (10 papers)
  2. Yinpei Dai (17 papers)
  3. Yinhe Zheng (30 papers)
  4. Yuchuan Wu (33 papers)
  5. Zheng Cao (48 papers)
  6. Dermot Liu (1 paper)
  7. Peng Jiang (272 papers)
  8. Min Yang (239 papers)
  9. Fei Huang (408 papers)
  10. Luo Si (73 papers)
  11. Jian Sun (414 papers)
  12. Yongbin Li (128 papers)
Citations (146)