Ziyang Luo Hong Kong Baptist University, Can Xu [email protected] Hong Kong Baptist University, Pu Zhao [email protected] Hong Kong Baptist University, Qingfeng Sun Hong Kong Baptist University, Xiubo Geng [email protected] Hong Kong Baptist University, Wenxiang Hu Hong Kong Baptist University, Chongyang Tao [email protected] Hong Kong Baptist University, Jing Ma [email protected] Hong Kong Baptist University, Qingwei Lin [email protected] Hong Kong Baptist University, Daxin Jiang [email protected] Hong Kong Baptist University, Microsoft Hong Kong Baptist University (2023)
The paper introduces WizardCoder, a Code Large Language Model (LLM) that leverages the Evol-Instruct method to enhance code generation capabilities through instruction fine-tuning. Unlike prior models that rely solely on extensive raw code data, WizardCoder focuses on creating complex instruction datasets specific to coding tasks. The authors conducted experiments on four code generation benchmarks: HumanEval, HumanEval+, MBPP, and DS-1000, demonstrating that WizardCoder outperforms existing open-source and even some closed-source models, including Claude and Bard. The paper details the methodology, experimental setup, and results, providing insights into the effectiveness of the Evol-Instruct adaptations for code-related tasks.
This paper employs the following methods:
The following datasets were used in this research:
The authors identified the following limitations: