Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Neural Text Generation: Past, Present and Beyond (1803.07133v1)

Published 15 Mar 2018 in cs.CL, cs.AI, and cs.LG

Abstract: This paper presents a systematic survey on recent development of neural text generation models. Specifically, we start from recurrent neural network LLMs with the traditional maximum likelihood estimation training scheme and point out its shortcoming for text generation. We thus introduce the recently proposed methods for text generation based on reinforcement learning, re-parametrization tricks and generative adversarial nets (GAN) techniques. We compare different properties of these models and the corresponding techniques to handle their common problems such as gradient vanishing and generation diversity. Finally, we conduct a benchmarking experiment with different types of neural text generation models on two well-known datasets and discuss the empirical results along with the aforementioned model properties.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Sidi Lu (13 papers)
  2. Yaoming Zhu (13 papers)
  3. Weinan Zhang (322 papers)
  4. Jun Wang (991 papers)
  5. Yong Yu (219 papers)
Citations (67)

Summary

We haven't generated a summary for this paper yet.