Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

CoT: Cooperative Training for Generative Modeling of Discrete Data (1804.03782v3)

Published 11 Apr 2018 in cs.LG, cs.AI, cs.CL, and stat.ML

Abstract: In this paper, we study the generative models of sequential discrete data. To tackle the exposure bias problem inherent in maximum likelihood estimation (MLE), generative adversarial networks (GANs) are introduced to penalize the unrealistic generated samples. To exploit the supervision signal from the discriminator, most previous models leverage REINFORCE to address the non-differentiable problem of sequential discrete data. However, because of the unstable property of the training signal during the dynamic process of adversarial training, the effectiveness of REINFORCE, in this case, is hardly guaranteed. To deal with such a problem, we propose a novel approach called Cooperative Training (CoT) to improve the training of sequence generative models. CoT transforms the min-max game of GANs into a joint maximization framework and manages to explicitly estimate and optimize Jensen-Shannon divergence. Moreover, CoT works without the necessity of pre-training via MLE, which is crucial to the success of previous methods. In the experiments, compared to existing state-of-the-art methods, CoT shows superior or at least competitive performance on sample quality, diversity, as well as training stability.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Sidi Lu (13 papers)
  2. Lantao Yu (32 papers)
  3. Siyuan Feng (55 papers)
  4. Yaoming Zhu (13 papers)
  5. Weinan Zhang (322 papers)
  6. Yong Yu (219 papers)
Citations (19)

Summary

We haven't generated a summary for this paper yet.