123-cao/Qwen2-0.5B-Instruct
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Mar 16, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

Qwen2-0.5B-Instruct is a 0.5 billion parameter instruction-tuned causal language model from the Qwen2 series, developed by Qwen. Built on a Transformer architecture with SwiGLU activation and group query attention, it features an improved tokenizer for multilingual and code adaptability. This model demonstrates competitive performance across benchmarks for language understanding, generation, coding, mathematics, and reasoning, making it suitable for a wide range of general-purpose conversational AI applications.

Loading preview...