Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Massive Supervised Fine-tuning Experiments Reveal How Data, Layer, and Training Factors Shape LLM Alignment Quality (2506.14681v1)

Published 17 Jun 2025 in cs.CL

Abstract: Supervised fine-tuning (SFT) is a critical step in aligning LLMs with human instructions and values, yet many aspects of SFT remain poorly understood. We trained a wide range of base models on a variety of datasets including code generation, mathematical reasoning, and general-domain tasks, resulting in 1,000+ SFT models under controlled conditions. We then identified the dataset properties that matter most and examined the layer-wise modifications introduced by SFT. Our findings reveal that some training-task synergies persist across all models while others vary substantially, emphasizing the importance of model-specific strategies. Moreover, we demonstrate that perplexity consistently predicts SFT effectiveness--often surpassing superficial similarity between trained data and benchmark--and that mid-layer weight changes correlate most strongly with performance gains. We will release these 1,000+ SFT models and benchmark results to accelerate further research.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yuto Harada (2 papers)
  2. Yusuke Yamauchi (4 papers)
  3. Yusuke Oda (15 papers)
  4. Yohei Oseki (22 papers)
  5. Yusuke Miyao (35 papers)
  6. Yu Takagi (11 papers)

Summary

We haven't generated a summary for this paper yet.