husheng12345/Qwen2.5-32B-Instruct
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Mar 16, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The husheng12345/Qwen2.5-32B-Instruct is a 32.5 billion parameter instruction-tuned causal language model developed by Qwen, featuring a transformer architecture with RoPE, SwiGLU, and RMSNorm. It offers significantly improved capabilities in coding, mathematics, and instruction following, alongside enhanced long text generation up to 8K tokens and structured output generation like JSON. This model supports a 131,072 token context length and is multilingual, covering over 29 languages, making it suitable for diverse complex language understanding and generation tasks.

Loading preview...