Jinze Bai Qwen Team Alibaba Group *, Shuai Bai Qwen Team Alibaba Group *, Yunfei Chu Qwen Team Alibaba Group *, Zeyu Cui Qwen Team Alibaba Group *, Kai Dang Qwen Team Alibaba Group *, Xiaodong Deng Qwen Team Alibaba Group *, Yang Fan Qwen Team Alibaba Group *, Wenbin Ge Qwen Team Alibaba Group *, Yu Han Qwen Team Alibaba Group *, Fei Huang Qwen Team Alibaba Group *, Binyuan Hui Qwen Team Alibaba Group *, Luo Ji Qwen Team Alibaba Group *, Mei Li Qwen Team Alibaba Group *, Junyang Lin Qwen Team Alibaba Group *, Runji Lin Qwen Team Alibaba Group *, Dayiheng Liu Qwen Team Alibaba Group *, Gao Liu Qwen Team Alibaba Group *, Chengqiang Lu Qwen Team Alibaba Group *, Keming Lu Qwen Team Alibaba Group *, Jianxin Ma Qwen Team Alibaba Group *, Rui Men Qwen Team Alibaba Group *, Xingzhang Ren Qwen Team Alibaba Group *, Xuancheng Ren Qwen Team Alibaba Group *, Chuanqi Tan Qwen Team Alibaba Group *, Sinan Tan Qwen Team Alibaba Group *, Jianhong Tu Qwen Team Alibaba Group *, Peng Wang Qwen Team Alibaba Group *, Shijie Wang Qwen Team Alibaba Group *, Wei Wang Qwen Team Alibaba Group *, Shengguang Wu Qwen Team Alibaba Group *, Benfeng Xu Qwen Team Alibaba Group *, Jin Xu Qwen Team Alibaba Group *, An Yang Qwen Team Alibaba Group *, Hao Yang Qwen Team Alibaba Group *, Jian Yang Qwen Team Alibaba Group *, Shusheng Yang Qwen Team Alibaba Group *, Yang Yao Qwen Team Alibaba Group *, Bowen Yu Qwen Team Alibaba Group *, Hongyi Yuan Qwen Team Alibaba Group *, Zheng Yuan Qwen Team Alibaba Group *, Jianwei Zhang Qwen Team Alibaba Group *, Xingxuan Zhang Qwen Team Alibaba Group *, Yichang Zhang Qwen Team Alibaba Group *, Zhenru Zhang Qwen Team Alibaba Group *, Chang Zhou Qwen Team Alibaba Group *, Jingren Zhou Qwen Team Alibaba Group *, Xiaohuan Zhou Qwen Team Alibaba Group *, Tianhang Zhu Qwen Team Alibaba Group * (2023)
This report presents QWEN, a series of large language models developed by Alibaba's Qwen Team. It introduces two main models: the base pretrained language models, QWEN, and the finetuned chat models, QWEN-CHAT, which use human alignment techniques for better performance on various language tasks. QWEN models have been pre-trained on 3 trillion tokens and utilize advanced training methodologies, including Reinforcement Learning from Human Feedback (RLHF). Specialized models such as CODE-QWEN and MATH-QWEN-CHAT are designed for coding and mathematics tasks. The evaluation demonstrates their competitive performance on benchmark datasets including HumanEval and GSM8K, with QWEN-CHAT outperforming smaller models and approaching proprietary solutions like GPT-4.
This paper employs the following methods:
The following datasets were used in this research:
The authors identified the following limitations: