DialogVED: A Pre-trained Latent Variable Encoder-Decoder Model for Dialog Response Generation (2204.13031v2)
Abstract: Dialog response generation in open domain is an important research topic where the main challenge is to generate relevant and diverse responses. In this paper, we propose a new dialog pre-training framework called DialogVED, which introduces continuous latent variables into the enhanced encoder-decoder pre-training framework to increase the relevance and diversity of responses. With the help of a large dialog corpus (Reddit), we pre-train the model using the following 4 tasks adopted in LLMs (LMs) and variational autoencoders (VAEs): 1) masked LLM; 2) response generation; 3) bag-of-words prediction; and 4) KL divergence reduction. We also add additional parameters to model the turn structure in dialogs to improve the performance of the pre-trained model. We conduct experiments on PersonaChat, DailyDialog, and DSTC7-AVSD benchmarks for response generation. Experimental results show that our model achieves the new state-of-the-art results on all these datasets.
- Wei Chen (1288 papers)
- Yeyun Gong (78 papers)
- Song Wang (313 papers)
- Bolun Yao (4 papers)
- Weizhen Qi (15 papers)
- Zhongyu Wei (98 papers)
- Xiaowu Hu (2 papers)
- Bartuer Zhou (4 papers)
- Yi Mao (78 papers)
- Weizhu Chen (128 papers)
- Biao Cheng (4 papers)
- Nan Duan (172 papers)