Skip to yearly menu bar Skip to main content


Poster

WizardLM: Empowering Large Pre-Trained Language Models to Follow Complex Instructions

Can Xu · Qingfeng Sun · Kai Zheng · Xiubo Geng · Pu Zhao · Jiazhan Feng · Chongyang Tao · Qingwei Lin · Daxin Jiang

Halle B #277

Abstract:

Training large language models (LLMs) with open-domain instruction following data brings colossal success. However, manually creating such instruction data is very time-consuming and labor-intensive. Moreover, humans may struggle to produce high-complexity instructions. In this paper, we show an avenue for creating large amounts of instruction data with varying levels of complexity using LLM instead of humans. Starting with an initial set of instructions, we use our proposed Evol-Instruct to rewrite them step by step into more complex instructions. Then, we mix all generated instruction data to fine-tune LLaMA. We call the resulting model WizardLM. Both automatic and human evaluations consistently indicate that WizardLM outperforms baselines such as Alpaca (trained from Self-Instruct) and Vicuna (trained from human-created instructions). The experimental results demonstrate that the quality of instruction-following dataset crafted by Evol-Instruct can significantly improve the performance of LLMs.

Chat is not available.