WSX/Qwen2.5-1.5B-Open-R1-GRPO-FC
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Feb 17, 2025Architecture:Transformer Warm

WSX/Qwen2.5-1.5B-Open-R1-GRPO-FC is a 1.5 billion parameter language model fine-tuned by WSX. It is based on the Qwen2.5 architecture and was trained using the GRPO method on the AI-MO/NuminaMath-TIR dataset. This model is specifically optimized for mathematical reasoning tasks, leveraging techniques from DeepSeekMath. Its primary strength lies in enhancing mathematical problem-solving capabilities within a compact parameter count.

Loading preview...