Can Xu [email protected] Peking University, Qingfeng Sun Peking University, Kai Zheng [email protected] Peking University, Xiubo Geng [email protected] Peking University, Pu Zhao [email protected] Peking University, Jiazhan Feng [email protected] Peking University, † Chongyang Tao [email protected] Peking University, Qingwei Lin [email protected] Peking University, Daxin Jiang [email protected] Peking University, Microsoft Peking University (2023)
This paper presents WizardLM, a large language model trained to follow complex instructions by generating diverse instruction data using a method called Evol-Instruct. The authors argue that creating open-domain instruction data using a language model can overcome the challenges of manually creating such data which is labor-intensive and often lacks complexity. The Evol-Instruct method rewrites initial simple instructions to more complex ones and generates a large dataset of 250,000 instructions to fine-tune LLaMA, resulting in WizardLM. Human evaluations indicate that WizardLM’s outputs are preferred over those from human-created datasets and even ChatGPT in complex scenarios, with great performance in GPT-4 automatic evaluations. Through this process, the authors demonstrate the potential of AI-evolved instructions for enhancing language model capabilities.
This paper employs the following methods:
The following datasets were used in this research:
The authors identified the following limitations: