ljwclass/Qwen2-0.5B-Instruct
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Mar 16, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

Qwen2-0.5B-Instruct is an instruction-tuned causal language model from the Qwen2 series developed by Qwen, featuring 0.5 billion parameters and a 32768-token context length. Built on a Transformer architecture with SwiGLU activation and group query attention, it demonstrates improved performance across language understanding, generation, multilingual capabilities, coding, mathematics, and reasoning benchmarks compared to its predecessor. This model is optimized for general-purpose instruction following and serves as a competitive option for applications requiring a compact yet capable LLM.

Loading preview...