Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Unified Pre-training Framework for Conversational AI (2105.02482v2)

Published 6 May 2021 in cs.CL

Abstract: In this work, we explore the application of PLATO-2 on various dialogue systems, including open-domain conversation, knowledge grounded dialogue, and task-oriented conversation. PLATO-2 is initially designed as an open-domain chatbot, trained via two-stage curriculum learning. In the first stage, a coarse-grained response generation model is learned to fit the simplified one-to-one mapping relationship. This model is applied to the task-oriented conversation, given that the semantic mappings tend to be deterministic in task completion. In the second stage, another fine-grained generation model and an evaluation model are further learned for diverse response generation and coherence estimation, respectively. With superior capability on capturing one-to-many mapping, such models are suitable for the open-domain conversation and knowledge grounded dialogue. For the comprehensive evaluation of PLATO-2, we have participated in multiple tasks of DSTC9, including interactive evaluation of open-domain conversation (Track3-task2), static evaluation of knowledge grounded dialogue (Track3-task1), and end-to-end task-oriented conversation (Track2-task1). PLATO-2 has obtained the 1st place in all three tasks, verifying its effectiveness as a unified framework for various dialogue systems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Siqi Bao (21 papers)
  2. Bingjin Chen (2 papers)
  3. Huang He (14 papers)
  4. Xin Tian (39 papers)
  5. Han Zhou (72 papers)
  6. Fan Wang (312 papers)
  7. Hua Wu (191 papers)
  8. Haifeng Wang (194 papers)
  9. Wenquan Wu (12 papers)
  10. Yingzhan Lin (6 papers)
Citations (4)