Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Task-oriented Dialog Model with Task-progressive and Policy-aware Pre-training (2310.00597v1)

Published 1 Oct 2023 in cs.CL

Abstract: Pre-trained conversation models (PCMs) have achieved promising progress in recent years. However, existing PCMs for Task-oriented dialog (TOD) are insufficient for capturing the sequential nature of the TOD-related tasks, as well as for learning dialog policy information. To alleviate these problems, this paper proposes a task-progressive PCM with two policy-aware pre-training tasks. The model is pre-trained through three stages where TOD-related tasks are progressively employed according to the task logic of the TOD system. A global policy consistency task is designed to capture the multi-turn dialog policy sequential relation, and an act-based contrastive learning task is designed to capture similarities among samples with the same dialog policy. Our model achieves better results on both MultiWOZ and In-Car end-to-end dialog modeling benchmarks with only 18\% parameters and 25\% pre-training data compared to the previous state-of-the-art PCM, GALAXY.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Lucen Zhong (3 papers)
  2. Hengtong Lu (1 paper)
  3. Caixia Yuan (13 papers)
  4. Xiaojie Wang (108 papers)
  5. Jiashen Sun (1 paper)
  6. Ke Zeng (15 papers)
  7. Guanglu Wan (24 papers)