Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

WizardLM: Empowering Large Language Models to Follow Complex Instructions (2304.12244v2)

Published 24 Apr 2023 in cs.CL and cs.AI
WizardLM: Empowering Large Language Models to Follow Complex Instructions

Abstract: Training LLMs with open-domain instruction following data brings colossal success. However, manually creating such instruction data is very time-consuming and labor-intensive. Moreover, humans may struggle to produce high-complexity instructions. In this paper, we show an avenue for creating large amounts of instruction data with varying levels of complexity using LLM instead of humans. Starting with an initial set of instructions, we use our proposed Evol-Instruct to rewrite them step by step into more complex instructions. Then, we mix all generated instruction data to fine-tune LLaMA. We call the resulting model WizardLM. Human evaluations on a complexity-balanced test bed and Vicuna's testset show that instructions from Evol-Instruct are superior to human-created ones. By analyzing the human evaluation results of the high complexity part, we demonstrate that outputs from our WizardLM are preferred to outputs from OpenAI ChatGPT. In GPT-4 automatic evaluation, WizardLM achieves more than 90\% capacity of ChatGPT on 17 out of 29 skills. Even though WizardLM still lags behind ChatGPT in some aspects, our findings suggest that fine-tuning with AI-evolved instructions is a promising direction for enhancing LLMs. Our code and data are public at https://github.com/nlpxucan/WizardLM

Overview of "WizardLM: Empowering LLMs to Follow Complex Instructions"

In the field of NLP, LLMs have proven adept at a variety of tasks, from generating coherent text to answering complex queries. However, their efficacy is often limited by their ability to follow user-specified instructions, particularly when those instructions are complex. The paper "WizardLM: Empowering LLMs to Follow Complex Instructions" addresses this limitation by introducing an innovative approach to generating complex instruction data for LLM training.

Methodology: Evol-Instruct

The crux of the paper's proposed method, termed Evol-Instruct, is the automated generation of complex instruction data using existing LLMs rather than human annotators, which is traditionally a resource-intensive process. The authors employ an iterative process that begins with a basic set of instructions. These instructions are then submitted to the Evol-Instruct system where they are progressively rewritten into more complex directives through a two-pronged approach: in-depth and in-breadth evolving.

  1. In-depth Evolving: This involves five specific operations designed to increase the complexity:
    • Adding constraints
    • Deepening (increasing the level of inquiry)
    • Concretizing (using more specific terms)
    • Increasing reasoning steps
    • Complicating input (adding complex input formats like code or data tables)
  2. In-breadth Evolving: This generates new instructions based on the themes of the initial set to diversify the instruction set's topic coverage.

A filtering mechanism, termed elimination evolving, validates the evolved instructions to ensure they are reasonable and not overly derived from evolving prompts.

Experimental Validation

The evolved instruction data is used to fine-tune the LLaMA LLM, resulting in a new model termed WizardLM. The efficacy of WizardLM is evaluated using a complexity-balanced test set and Vicuna's test set, both through human assessment and automatic scoring using GPT-4. The results show that WizardLM outperforms existing models, including OpenAI's ChatGPT and other instruction-tuned models like Vicuna and Alpaca, particularly on complex instructional tasks.

  1. Human Evaluations: On the Evol-Instruct and Vicuna test sets, WizardLM displayed superior performance compared to Vicuna and Alpaca, with notable improvement in handling high-difficulty instructions.
  2. GPT-4 Automatic Evaluation: WizardLM achieved over 90% of ChatGPT’s capacity in 17 out of 29 skills, marking significant advancements in instruction following capabilities through automated data generation.

Implications and Future Work

The implications of this research are profound for both practical applications and theoretical advancements. Practically, the ability to generate complex instructional data autonomously can significantly reduce the cost and time associated with human annotation, making it more feasible to train LLMs to comprehend and execute more complex tasks.

Theoretically, the findings indicate that AI-generated data can match, if not exceed, the quality and complexity of human-generated data, thus suggesting a potential paradigm shift in how training datasets for LLMs are constructed. The authors suggest that future work could focus on refining the elimination evolving process and exploring the integration of more sophisticated evolving techniques to further enhance the complexity and application range of the instructions.

Conclusion

The paper "WizardLM: Empowering LLMs to Follow Complex Instructions" presents a significant advancement in the field of NLP by demonstrating how LLMs can be fine-tuned with automatically generated complex instructions. This method not only enhances the capabilities of LLMs to follow intricate user directives but also offers a sustainable and scalable approach to data generation that could revolutionize training methodologies for sophisticated language tasks. The introduction of the Evol-Instruct framework sets a promising direction for future research and application in autonomous AI instruction and comprehension, potentially leading to more intelligent and adaptable AI systems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Can Xu (98 papers)
  2. Qingfeng Sun (40 papers)
  3. Kai Zheng (134 papers)
  4. Xiubo Geng (36 papers)
  5. Pu Zhao (82 papers)
  6. Jiazhan Feng (11 papers)
  7. Chongyang Tao (61 papers)
  8. Daxin Jiang (138 papers)
Citations (769)
Github Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com