WizardCoder: Empowering Code Large Language Models with Evol-Instruct

Article Status
Published
Authors/contributors
Title
WizardCoder: Empowering Code Large Language Models with Evol-Instruct
Abstract
Code Large Language Models (Code LLMs), such as StarCoder, have demonstrated exceptional performance in code-related tasks. However, most existing models are solely pre-trained on extensive raw code data without instruction fine-tuning. In this paper, we introduce WizardCoder, which empowers Code LLMs with complex instruction fine-tuning, by adapting the Evol-Instruct method to the domain of code. Through comprehensive experiments on four prominent code generation benchmarks, namely HumanEval, HumanEval+, MBPP, and DS-1000, we unveil the exceptional capabilities of our model. It surpasses all other open-source Code LLMs by a substantial margin. Moreover, our model even outperforms the largest closed LLMs, Anthropic's Claude and Google's Bard, on HumanEval and HumanEval+. Our code, model weights, and data are public at https://github.com/nlpxucan/WizardLM
Repository
arXiv
Archive ID
arXiv:2306.08568
Date
2023-06-14
Accessed
01/09/2023, 12:23
Short Title
WizardCoder
Library Catalogue
Extra
arXiv:2306.08568 [cs]
Citation
Luo, Z., Xu, C., Zhao, P., Sun, Q., Geng, X., Hu, W., Tao, C., Ma, J., Lin, Q., & Jiang, D. (2023). WizardCoder: Empowering Code Large Language Models with Evol-Instruct (arXiv:2306.08568). arXiv. http://arxiv.org/abs/2306.08568
Powered by Zotero and Kerko.