Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

AlchemistCoder: Harmonizing and Eliciting Code Capability by Hindsight Tuning on Multi-source Data (2405.19265v1)

Published 29 May 2024 in cs.CL

Abstract: Open-source LLMs and their specialized variants, particularly Code LLMs, have recently delivered impressive performance. However, previous Code LLMs are typically fine-tuned on single-source data with limited quality and diversity, which may insufficiently elicit the potential of pre-trained Code LLMs. In this paper, we present AlchemistCoder, a series of Code LLMs with enhanced code generation and generalization capabilities fine-tuned on multi-source data. To achieve this, we pioneer to unveil inherent conflicts among the various styles and qualities in multi-source code corpora and introduce data-specific prompts with hindsight relabeling, termed AlchemistPrompts, to harmonize different data sources and instruction-response pairs. Additionally, we propose incorporating the data construction process into the fine-tuning data as code comprehension tasks, including instruction evolution, data filtering, and code review. Extensive experiments demonstrate that AlchemistCoder holds a clear lead among all models of the same size (6.7B/7B) and rivals or even surpasses larger models (15B/33B/70B), showcasing the efficacy of our method in refining instruction-following capabilities and advancing the boundaries of code intelligence.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Zifan Song (5 papers)
  2. Yudong Wang (28 papers)
  3. Wenwei Zhang (77 papers)
  4. Kuikun Liu (12 papers)
  5. Chengqi Lyu (13 papers)
  6. Demin Song (11 papers)
  7. Qipeng Guo (72 papers)
  8. Hang Yan (86 papers)
  9. Dahua Lin (336 papers)
  10. Kai Chen (512 papers)
  11. Cairong Zhao (24 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets