Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Semi-Supervised Knowledge-Grounded Pre-training for Task-Oriented Dialog Systems (2210.08873v2)

Published 17 Oct 2022 in cs.CL

Abstract: Recent advances in neural approaches greatly improve task-oriented dialogue (TOD) systems which assist users to accomplish their goals. However, such systems rely on costly manually labeled dialogs which are not available in practical scenarios. In this paper, we present our models for Track 2 of the SereTOD 2022 challenge, which is the first challenge of building semi-supervised and reinforced TOD systems on a large-scale real-world Chinese TOD dataset MobileCS. We build a knowledge-grounded dialog model to formulate dialog history and local KB as input and predict the system response. And we perform semi-supervised pre-training both on the labeled and unlabeled data. Our system achieves the first place both in the automatic evaluation and human interaction, especially with higher BLEU (+7.64) and Success (+13.6\%) than the second place.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Weihao Zeng (24 papers)
  2. Keqing He (47 papers)
  3. Zechen Wang (15 papers)
  4. Dayuan Fu (13 papers)
  5. Guanting Dong (46 papers)
  6. Ruotong Geng (3 papers)
  7. Pei Wang (240 papers)
  8. Jingang Wang (71 papers)
  9. Chaobo Sun (4 papers)
  10. Wei Wu (481 papers)
  11. Weiran Xu (58 papers)
Citations (15)