Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

PAFT: A Parallel Training Paradigm for Effective LLM Fine-Tuning (2406.17923v1)

Published 25 Jun 2024 in cs.CL
PAFT: A Parallel Training Paradigm for Effective LLM Fine-Tuning

Abstract: LLMs have shown remarkable abilities in diverse NLP tasks. The LLMs generally undergo supervised fine-tuning (SFT) followed by preference alignment to be usable in downstream applications. However, this sequential training pipeline leads to alignment tax that degrades the LLM performance. This paper introduces PAFT, a new PArallel training paradigm for effective LLM Fine-Tuning, which independently performs SFT and preference alignment (e.g., DPO and ORPO, etc.) with the same pre-trained model on respective datasets. The model produced by SFT and the model from preference alignment are then merged into a final model by parameter fusing for use in downstream applications. This work reveals important findings that preference alignment like DPO naturally results in a sparse model while SFT leads to a natural dense model which needs to be sparsified for effective model merging. This paper introduces an effective interference resolution which reduces the redundancy by sparsifying the delta parameters. The LLM resulted from the new training paradigm achieved Rank #1 on the HuggingFace Open LLM Leaderboard. Comprehensive evaluation shows the effectiveness of the parallel training paradigm.

Evaluation of PAFT: A Parallel Training Paradigm for Effective LLM Fine-Tuning

The paper "PAFT: A Parallel Training Paradigm for Effective LLM Fine-Tuning" addresses a fundamental challenge in the application of LLMs in NLP tasks, namely the "alignment tax" that arises from the sequential nature of existing fine-tuning frameworks. The typical approach of employing supervised fine-tuning (SFT), followed by a preference alignment stage often results in performance degradation due to catastrophic forgetting. This work proposes a novel parallel training paradigm (PAFT) that concurrently performs SFT and preference alignment, significantly mitigating the detrimental effects associated with sequential workflows.

Methodology and Results

The PAFT model architecture deviates from traditional sequential fine-tuning paradigms by concurrently applying SFT and preference alignment algorithms, like Direct Preference Optimization (DPO) or Odds-ratio Preference Optimization (ORPO), on the same pre-trained model across distinct datasets. The primary insight here is that the model emerges from SFT as inherently dense, leading to complications during the merging process with the preference-aligned model. PAFT introduces an L1-norm penalty during the SFT stage to induce sparsity in delta parameters, effectively reducing interference when merging models.

Comprehensive experiments underscore the efficacy of PAFT. Notably, models fine-tuned under the PAFT regime achieve notable rankings in public benchmarks, such as the HuggingFace Open LLM Leaderboard. PAFT consistently outperformed both independent and sequentially fine-tuned models. The integration of the L1-norm penalty enhanced model sparsity and facilitated more effective parameter merging, as evidenced by the concrete benchmark results.

Implications

This paper posits several significant implications for the future of NLP application and model training. Firstly, by providing a robust method for overcoming the alignment tax associated with sequential training, PAFT offers a pathway towards more refined and efficient LLM deployment across diverse tasks. This heightened efficiency will likely inspire further research into parallel processing paradigms and sparsity techniques in LLMs.

Secondly, the explicit inclusion of sparsity in the model fine-tuning process holds substantial promise for enhancing model performance post-merging. The findings encouraging the use of parallel training paradigms enable fully leveraging large-scale data and diverse task criteria, potentially setting a new standard for LLM fine-tuning methodologies.

Future Directions

This research suggests several pathways for future exploration. Given the success of PAFT with DPO and ORPO, there lies an opportunity to evaluate the applicability of PAFT with alternative alignment algorithms, potentially broadening its applicability and robustness across a wider array of NLP tasks.

Further studies could also examine the theoretical underpinnings of model sparsity's role in reducing training conflicts during parallel processing. By rigorously evaluating the balance between sparsity and model performance, researchers can better comprehend the dynamics behind improved model coherence, leading to more sophisticated fine-tuning methods.

In conclusion, the PAFT framework offers a promising strategy for overcoming sequential fine-tuning limitations by leveraging parallel processing and sparsity. These advancements signal a shift towards more efficient and powerful LLM fine-tuning strategies, opening avenues for continued innovation in NLP model training methodologies.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Shiva Kumar Pentyala (4 papers)
  2. Zhichao Wang (83 papers)
  3. Bin Bi (24 papers)
  4. Kiran Ramnath (4 papers)
  5. Xiang-Bo Mao (3 papers)
  6. Regunathan Radhakrishnan (2 papers)
  7. Sitaram Asur (13 papers)
  8. Na (2 papers)
  9. Cheng (8 papers)
Citations (6)