Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ARAML: A Stable Adversarial Training Framework for Text Generation (1908.07195v1)

Published 20 Aug 2019 in cs.CL and cs.LG

Abstract: Most of the existing generative adversarial networks (GAN) for text generation suffer from the instability of reinforcement learning training algorithms such as policy gradient, leading to unstable performance. To tackle this problem, we propose a novel framework called Adversarial Reward Augmented Maximum Likelihood (ARAML). During adversarial training, the discriminator assigns rewards to samples which are acquired from a stationary distribution near the data rather than the generator's distribution. The generator is optimized with maximum likelihood estimation augmented by the discriminator's rewards instead of policy gradient. Experiments show that our model can outperform state-of-the-art text GANs with a more stable training process.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Pei Ke (37 papers)
  2. Fei Huang (408 papers)
  3. Minlie Huang (225 papers)
  4. Xiaoyan Zhu (54 papers)
Citations (22)