Kazuki1450/Qwen3-0.6B_geo_3_6_clean_1p0_0p0_1p0_grpo_42_rule
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Mar 16, 2026Architecture:Transformer Cold

Kazuki1450/Qwen3-0.6B_geo_3_6_clean_1p0_0p0_1p0_grpo_42_rule is an 0.8 billion parameter language model fine-tuned from Qwen/Qwen3-0.6B. This model was trained using the GRPO method, which is designed to enhance mathematical reasoning capabilities. It is specifically optimized for tasks requiring advanced mathematical problem-solving, leveraging techniques from the DeepSeekMath research. The model is suitable for applications where robust mathematical understanding and generation are critical.

Loading preview...