wuyifan666/Qwen2.5-0.5B
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Mar 31, 2026License:apache-2.0Architecture:Transformer Open Weights Loading

The wuyifan666/Qwen2.5-0.5B is a 0.49 billion parameter causal language model from the Qwen2.5 series, developed by Qwen. This base model features a transformer architecture with RoPE, SwiGLU, and RMSNorm, supporting a 32,768 token context length. It offers significant improvements in knowledge, coding, and mathematics compared to its predecessor, alongside enhanced instruction following and structured data understanding. It is designed for pretraining and is not recommended for direct conversational use.

Loading preview...