chenyukun/qwen3-0.6b-grpo-math
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Mar 13, 2026Architecture:Transformer Warm
The chenyukun/qwen3-0.6b-grpo-math model is a fine-tuned 0.8 billion parameter Qwen3-0.6B causal language model, developed by chenyukun, with a context length of 32768 tokens. It has been specifically trained using the GRPO method, as introduced in the DeepSeekMath paper, to enhance its mathematical reasoning capabilities. This model is optimized for tasks requiring robust mathematical problem-solving and logical deduction.
Loading preview...