Kazuki1450/Qwen3-1.7B-Base_dsum_3_6_0p5_0p0_1p0_grpo_sapo_42_rule
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 26, 2026Architecture:Transformer Warm

Kazuki1450/Qwen3-1.7B-Base_dsum_3_6_0p5_0p0_1p0_grpo_sapo_42_rule is a 2 billion parameter language model fine-tuned from Qwen/Qwen3-1.7B-Base. This model was trained using the GRPO method, as introduced in the DeepSeekMath paper, to enhance mathematical reasoning capabilities. With a context length of 32768 tokens, it is optimized for tasks requiring robust logical and mathematical processing. It is suitable for applications where a smaller model with strong reasoning skills is beneficial.

Loading preview...