felixluo1002/Qwen2.5-1.5B
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Loading

Qwen2.5-1.5B is a 1.54 billion parameter causal language model from the Qwen2.5 series, developed by Qwen Team. This base model features a 32,768 token context length and is designed with a transformer architecture including RoPE, SwiGLU, and RMSNorm. It offers significantly improved capabilities in coding, mathematics, instruction following, and long text generation, with multilingual support for over 29 languages.

Loading preview...