- Published on
Alibaba Cloud’s QwenLM has released Qwen2, the latest iteration of its language model series.
This release includes five models of varying sizes, from Qwen2-0.5B to Qwen2-72B, all of which have base and chat versions.
The models have been trained on data in 27 additional languages beyond English and Chinese, significantly expanding their multilingual capabilities. Qwen2 demonstrates state-of-the-art performance in a wide range of benchmark evaluations, with notable improvements in coding and mathematics.
The models also support extended context lengths, up to 128K tokens for the Qwen2-7B-Instruct and Qwen2-72B-Instruct models.
The Qwen2-57B-A14B version is a mixture of experts with 14B parameters active.
The 72B versions are released under the original Qianwen License, while all other models have adopted theĀ Apache 2.0 license.