vkasera/v3_qwen-2.5-3b-r1-countdown-phil
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Oct 3, 2025Architecture:Transformer Cold
The vkasera/v3_qwen-2.5-3b-r1-countdown-phil model is a 3.1 billion parameter language model fine-tuned from Qwen/Qwen2.5-3B-Instruct. Developed by vkasera, this model was trained using the GRPO method, as introduced in the DeepSeekMath paper, to enhance its reasoning capabilities. It is optimized for tasks requiring advanced reasoning, building upon the base Qwen2.5 architecture.
Loading preview...