Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Chain-of-Instructions: Compositional Instruction Tuning on Large Language Models (2402.11532v2)

Published 18 Feb 2024 in cs.CL

Abstract: Fine-tuning LLMs with a collection of large and diverse instructions has improved the model's generalization to different tasks, even for unseen tasks. However, most existing instruction datasets include only single instructions, and they struggle to follow complex instructions composed of multiple subtasks. In this work, we propose a novel concept of compositional instructions called chain-of-instructions (CoI), where the output of one instruction becomes an input for the next like a chain. Unlike the conventional practice of solving single instruction tasks, our proposed method encourages a model to solve each subtask step by step until the final answer is reached. CoI-tuning (i.e., fine-tuning with CoI instructions) improves the model's ability to handle instructions composed of multiple subtasks as well as unseen composite tasks such as multilingual summarization. Overall, our study find that simple CoI tuning of existing instruction data can provide consistent generalization to solve more complex, unseen, and longer chains of instructions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Shirley Anugrah Hayati (13 papers)
  2. Taehee Jung (4 papers)
  3. Tristan Bodding-Long (1 paper)
  4. Sudipta Kar (19 papers)
  5. Abhinav Sethy (14 papers)
  6. Joo-Kyung Kim (12 papers)
  7. Dongyeop Kang (72 papers)
Citations (5)