Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Survey on Data Selection for LLM Instruction Tuning (2402.05123v1)

Published 4 Feb 2024 in cs.CL

Abstract: Instruction tuning is a vital step of training LLMs (LLM), so how to enhance the effect of instruction tuning has received increased attention. Existing works indicate that the quality of the dataset is more crucial than the quantity during instruction tuning of LLM. Therefore, recently a lot of studies focus on exploring the methods of selecting high-quality subset from instruction datasets, aiming to reduce training costs and enhance the instruction-following capabilities of LLMs. This paper presents a comprehensive survey on data selection for LLM instruction tuning. Firstly, we introduce the wildly used instruction datasets. Then, we propose a new taxonomy of the data selection methods and provide a detailed introduction of recent advances,and the evaluation strategies and results of data selection methods are also elaborated in detail. Finally, we emphasize the open challenges and present new frontiers of this task.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jiahao Wang (88 papers)
  2. Bolin Zhang (10 papers)
  3. Qianlong Du (6 papers)
  4. Jiajun Zhang (176 papers)
  5. Dianhui Chu (14 papers)
Citations (28)
X Twitter Logo Streamline Icon: https://streamlinehq.com