Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Text Generation with Exemplar-based Adaptive Decoding (1904.04428v2)

Published 9 Apr 2019 in cs.CL

Abstract: We propose a novel conditioned text generation model. It draws inspiration from traditional template-based text generation techniques, where the source provides the content (i.e., what to say), and the template influences how to say it. Building on the successful encoder-decoder paradigm, it first encodes the content representation from the given input text; to produce the output, it retrieves exemplar text from the training data as "soft templates," which are then used to construct an exemplar-specific decoder. We evaluate the proposed model on abstractive text summarization and data-to-text generation. Empirical results show that this model achieves strong performance and outperforms comparable baselines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Hao Peng (291 papers)
  2. Ankur P. Parikh (28 papers)
  3. Manaal Faruqui (39 papers)
  4. Bhuwan Dhingra (66 papers)
  5. Dipanjan Das (42 papers)
Citations (56)