Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

End-to-End Content and Plan Selection for Data-to-Text Generation (1810.04700v1)

Published 10 Oct 2018 in cs.CL and cs.AI

Abstract: Learning to generate fluent natural language from structured data with neural networks has become an common approach for NLG. This problem can be challenging when the form of the structured data varies between examples. This paper presents a survey of several extensions to sequence-to-sequence models to account for the latent content selection process, particularly variants of copy attention and coverage decoding. We further propose a training method based on diverse ensembling to encourage models to learn distinct sentence templates during training. An empirical evaluation of these techniques shows an increase in the quality of generated text across five automated metrics, as well as human evaluation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Sebastian Gehrmann (48 papers)
  2. Falcon Z. Dai (9 papers)
  3. Henry Elder (4 papers)
  4. Alexander M. Rush (115 papers)
Citations (69)

Summary

We haven't generated a summary for this paper yet.