Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

From Symbolic Tasks to Code Generation: Diversification Yields Better Task Performers (2405.19787v2)

Published 30 May 2024 in cs.CL, cs.AI, cs.LG, cs.LO, and cs.PL

Abstract: Instruction tuning -- tuning LLMs on instruction-output pairs -- is a promising technique for making models better adapted to the real world. Yet, the key factors driving the model's capability to understand and follow instructions not seen during training remain under-explored. Our investigation begins with a series of synthetic experiments within the theoretical framework of a Turing-complete algorithm called Markov algorithm, which allows fine-grained control over the instruction-tuning data. Generalization and robustness with respect to the training distribution emerge once a diverse enough set of tasks is provided, even though very few examples are provided for each task. We extend these initial results to a real-world application scenario of code generation and find that a more diverse instruction set, extending beyond code-related tasks, improves the performance of code generation. Our observations suggest that a more diverse semantic space for instruction-tuning sets greatly improves the model's ability to follow instructions and perform tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Dylan Zhang (12 papers)
  2. Justin Wang (14 papers)
  3. Francois Charton (10 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com