The provided text is a technical report introducing the Qwen2 series, the latest suite of open-weight large language and multimodal models developed by the Qwen Team at Alibaba Group.
Key highlights from the report include:
- Model Variants: The Qwen2 release encompasses dense models with parameter sizes of 0.5 billion, 1.5 billion, 7 billion, and 72 billion, as well as a 57 billion parameter Mixture-of-Experts (MoE) model.
- Massive Training Data & Multilingualism: The models were pre-trained on a high-quality dataset of over 7 trillion tokens (and 12 trillion for the 0.5B model). As a result, Qwen2 demonstrates robust multilingual capabilities, with proficiency in approximately 30 languages including English, Chinese, Spanish, French, and Arabic.
- Architectural Advancements: The models incorporate structural upgrades such as Grouped Query Attention (GQA) for optimized inference and Dual Chunk Attention (DCA) combined with YARN. These enhancements allow the models to successfully process long-context scenarios of up to 128,000 tokens.
- State-of-the-Art Performance: Through rigorous evaluation, Qwen2—particularly its flagship Qwen2-72B model—outperforms its predecessor (Qwen1.5) and exhibits highly competitive performance against leading open-weight and proprietary models. It demonstrates notable strength across benchmarks in language understanding, coding, mathematics, and logical reasoning.
- Open Accessibility: To encourage community research and innovation, the Qwen2 model weights, alongside code and supplementary materials, have been made openly available on platforms such as Hugging Face, ModelScope, and GitHub.