Qwen2.5-7B-Instruct Overview
Qwen2.5-7B-Instruct is an instruction-tuned causal language model from the Qwen2.5 series, featuring 7.61 billion parameters. It builds upon the Qwen2 architecture with substantial enhancements in several key areas. The model incorporates transformers with RoPE, SwiGLU, RMSNorm, and Attention QKV bias, and supports a full context length of 131,072 tokens, with generation capabilities up to 8,192 tokens.
Key Capabilities & Improvements
- Enhanced Knowledge & Reasoning: Significantly improved capabilities in coding and mathematics, leveraging specialized expert models.
- Instruction Following & Generation: Offers better instruction following, improved generation of long texts (over 8K tokens), and enhanced understanding of structured data like tables.
- Structured Output: Excels at generating structured outputs, particularly JSON, and is more resilient to diverse system prompts for robust role-play and chatbot implementations.
- Long-Context Support: Features a full context length of 131,072 tokens, with a generation capacity of 8,192 tokens. It utilizes YaRN for optimal performance on lengthy texts, though static YaRN in vLLM may impact shorter text performance.
- Multilingual Support: Provides comprehensive support for over 29 languages, including Chinese, English, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, and Arabic.
Architecture & Training
This model is a causal language model that underwent both pretraining and post-training stages. It consists of 28 layers and uses 28 attention heads for Q and 4 for KV in its Grouped-query Attention (GQA) mechanism. For deployment with long texts, users can configure rope_scaling with "type": "yarn" in config.json.