Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Multimodal Response Generation with Exemplar Augmentation and Curriculum Optimization (2004.12429v1)

Published 26 Apr 2020 in cs.CL

Abstract: Recently, variational auto-encoder (VAE) based approaches have made impressive progress on improving the diversity of generated responses. However, these methods usually suffer the cost of decreased relevance accompanied by diversity improvements. In this paper, we propose a novel multimodal response generation framework with exemplar augmentation and curriculum optimization to enhance relevance and diversity of generated responses. First, unlike existing VAE-based models that usually approximate a simple Gaussian posterior distribution, we present a Gaussian mixture posterior distribution (i.e, multimodal) to further boost response diversity, which helps capture complex semantics of responses. Then, to ensure that relevance does not decrease while diversity increases, we fully exploit similar examples (exemplars) retrieved from the training data into posterior distribution modeling to augment response relevance. Furthermore, to facilitate the convergence of Gaussian mixture prior and posterior distributions, we devise a curriculum optimization strategy to progressively train the model under multiple training criteria from easy to hard. Experimental results on widely used SwitchBoard and DailyDialog datasets demonstrate that our model achieves significant improvements compared to strong baselines in terms of diversity and relevance.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Zeyang Lei (6 papers)
  2. Zekang Li (13 papers)
  3. Jinchao Zhang (49 papers)
  4. Fandong Meng (174 papers)
  5. Yang Feng (230 papers)
  6. Yujiu Yang (155 papers)
  7. Cheng Niu (15 papers)
  8. Jie Zhou (687 papers)

Summary

We haven't generated a summary for this paper yet.