f0rc3ps/Qwen2.5-3B
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026License:qwen-researchArchitecture:Transformer Cold

The f0rc3ps/Qwen2.5-3B model is a 3.09 billion parameter causal language model developed by Qwen, part of the Qwen2.5 series. This base model features a 32,768-token context length and is designed for pretraining, offering significant improvements in knowledge, coding, and mathematics compared to its predecessors. It excels in instruction following, long text generation, structured data understanding, and multilingual support across 29 languages, making it suitable for further fine-tuning for specialized applications.

Loading preview...