Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

To Beam Or Not To Beam: That is a Question of Cooperation for Language GANs (2106.06363v1)

Published 11 Jun 2021 in cs.CL and cs.AI

Abstract: Due to the discrete nature of words, language GANs require to be optimized from rewards provided by discriminator networks, via reinforcement learning methods. This is a much harder setting than for continuous tasks, which enjoy gradient flows from discriminators to generators, usually leading to dramatic learning instabilities. However, we claim that this can be solved by making discriminator and generator networks cooperate to produce output sequences during training. These cooperative outputs, inherently built to obtain higher discrimination scores, not only provide denser rewards for training, but also form a more compact artificial set for discriminator training, hence improving its accuracy and stability. In this paper, we show that our SelfGAN framework, built on this cooperative principle, outperforms Teacher Forcing and obtains state-of-the-art results on two challenging tasks, Summarization and Question Generation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Thomas Scialom (35 papers)
  2. Paul-Alexis Dray (7 papers)
  3. Sylvain Lamprier (40 papers)
  4. Benjamin Piwowarski (38 papers)
  5. Jacopo Staiano (38 papers)
Citations (17)

Summary

We haven't generated a summary for this paper yet.