aaasdsdfefsdfe/Qwen2.5-7B-Instruct
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Mar 22, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

Qwen2.5-7B-Instruct is a 7.61 billion parameter instruction-tuned causal language model developed by Qwen, featuring a transformer architecture with RoPE, SwiGLU, RMSNorm, and Attention QKV bias. It offers significant improvements in coding, mathematics, instruction following, and long text generation up to 8K tokens, with a context length of 131,072 tokens. This model excels at understanding structured data and generating structured outputs like JSON, and provides robust multilingual support for over 29 languages.

Loading preview...