Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Controlled Text Generation with Natural Language Instructions (2304.14293v2)

Published 27 Apr 2023 in cs.CL, cs.AI, and cs.LG

Abstract: LLMs generate fluent texts and can follow natural language instructions to solve a wide range of tasks without task-specific training. Nevertheless, it is notoriously difficult to control their generation to satisfy the various constraints required by different applications. In this work, we present InstructCTG, a controlled text generation framework that incorporates different constraints by conditioning on natural language descriptions and demonstrations of the constraints. In particular, we first extract the underlying constraints of natural texts through a combination of off-the-shelf NLP tools and simple heuristics. We then verbalize the constraints into natural language instructions to form weakly supervised training data. By prepending natural language descriptions of the constraints and a few demonstrations, we fine-tune a pre-trained LLM to incorporate various types of constraints. Compared to existing search-based or score-based methods, InstructCTG is more flexible to different constraint types and has a much smaller impact on the generation quality and speed because it does not modify the decoding procedure. Additionally, InstructCTG allows the model to adapt to new constraints without re-training through the use of few-shot task generalization and in-context learning abilities of instruction-tuned LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Wangchunshu Zhou (73 papers)
  2. Yuchen Eleanor Jiang (19 papers)
  3. Ethan Wilcox (24 papers)
  4. Ryan Cotterell (226 papers)
  5. Mrinmaya Sachan (124 papers)
Citations (68)

Summary

We haven't generated a summary for this paper yet.