Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Open-domain Implicit Format Control for Large Language Model Generation (2408.04392v1)

Published 8 Aug 2024 in cs.CL

Abstract: Controlling the format of outputs generated by LLMs is a critical functionality in various applications. Current methods typically employ constrained decoding with rule-based automata or fine-tuning with manually crafted format instructions, both of which struggle with open-domain format requirements. To address this limitation, we introduce a novel framework for controlled generation in LLMs, leveraging user-provided, one-shot QA pairs. This study investigates LLMs' capabilities to follow open-domain, one-shot constraints and replicate the format of the example answers. We observe that this is a non-trivial problem for current LLMs. We also develop a dataset collection methodology for supervised fine-tuning that enhances the open-domain format control of LLMs without degrading output quality, as well as a benchmark on which we evaluate both the helpfulness and format correctness of LLM outputs. The resulting datasets, named OIFC-SFT, along with the related code, will be made publicly available at https://github.com/cofe-ai/OIFC.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Yiqun Yao (14 papers)
  2. Wenjia Ma (2 papers)
  3. Xuezhi Fang (11 papers)
  4. Xin Jiang (242 papers)
  5. Xiang Li (1003 papers)
  6. Xuying Meng (18 papers)
  7. Peng Han (37 papers)
  8. Jing Li (621 papers)
  9. Aixin Sun (99 papers)
  10. Yequan Wang (44 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com