Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DFM: Dialogue Foundation Model for Universal Large-Scale Dialogue-Oriented Task Learning (2205.12662v2)

Published 25 May 2022 in cs.CL

Abstract: Building a universal conversational agent has been a long-standing goal of the dialogue research community. Most previous works only focus on a small set of dialogue tasks. In this work, we aim to build a unified dialogue foundation model (DFM) which can be used to solve massive diverse dialogue tasks. To achieve this goal, a large-scale well-annotated dialogue dataset with rich task diversity (DialogZoo) is collected. We introduce a framework to unify all dialogue tasks and propose novel auxiliary self-supervised tasks to achieve stable training of DFM on the highly diverse large scale DialogZoo corpus. Experiments show that, compared with models of the same size, DFM can achieve state-of-the-art or competitive performance on very rich cross-domain downstream dialogue tasks. This demonstrates that DFM largely extends the ability of unified dialogue pre-trained model.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (13)
  1. Zhi Chen (235 papers)
  2. Jijia Bao (1 paper)
  3. Lu Chen (244 papers)
  4. Yuncong Liu (7 papers)
  5. Da Ma (28 papers)
  6. Bei Chen (56 papers)
  7. Mengyue Wu (57 papers)
  8. Su Zhu (29 papers)
  9. Xin Dong (90 papers)
  10. Fujiang Ge (1 paper)
  11. Qingliang Miao (2 papers)
  12. Jian-Guang Lou (69 papers)
  13. Kai Yu (201 papers)
Citations (3)