Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Well-Composed Text is Half Done! Composition Sampling for Diverse Conditional Generation (2203.15108v1)

Published 28 Mar 2022 in cs.CL

Abstract: We propose Composition Sampling, a simple but effective method to generate diverse outputs for conditional generation of higher quality compared to previous stochastic decoding strategies. It builds on recently proposed plan-based neural generation models (Narayan et al, 2021) that are trained to first create a composition of the output and then generate by conditioning on it and the input. Our approach avoids text degeneration by first sampling a composition in the form of an entity chain and then using beam search to generate the best possible text grounded to this entity chain. Experiments on summarization (CNN/DailyMail and XSum) and question generation (SQuAD), using existing and newly proposed automatic metrics together with human-based evaluation, demonstrate that Composition Sampling is currently the best available decoding strategy for generating diverse meaningful outputs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Shashi Narayan (35 papers)
  2. Gonçalo Simões (7 papers)
  3. Yao Zhao (272 papers)
  4. Joshua Maynez (28 papers)
  5. Dipanjan Das (42 papers)
  6. Michael Collins (46 papers)
  7. Mirella Lapata (135 papers)
Citations (29)