Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Controlled Text Generation as Continuous Optimization with Multiple Constraints (2108.01850v1)

Published 4 Aug 2021 in cs.CL

Abstract: As large-scale LLM pretraining pushes the state-of-the-art in text generation, recent work has turned to controlling attributes of the text such models generate. While modifying the pretrained models via fine-tuning remains the popular approach, it incurs a significant computational cost and can be infeasible due to lack of appropriate data. As an alternative, we propose MuCoCO -- a flexible and modular algorithm for controllable inference from pretrained models. We formulate the decoding process as an optimization problem which allows for multiple attributes we aim to control to be easily incorporated as differentiable constraints to the optimization. By relaxing this discrete optimization to a continuous one, we make use of Lagrangian multipliers and gradient-descent based techniques to generate the desired text. We evaluate our approach on controllable machine translation and style transfer with multiple sentence-level attributes and observe significant improvements over baselines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Sachin Kumar (68 papers)
  2. Eric Malmi (26 papers)
  3. Aliaksei Severyn (29 papers)
  4. Yulia Tsvetkov (142 papers)
Citations (68)

Summary

We haven't generated a summary for this paper yet.