Kazuki1450/Qwen3-1.7B-Base_csum_3_10_tok_parentheses_1p0_0p0_1p0_grpo_42_rule
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 18, 2026Architecture:Transformer Loading
Kazuki1450/Qwen3-1.7B-Base_csum_3_10_tok_parentheses_1p0_0p0_1p0_grpo_42_rule is a 1.7 billion parameter language model, fine-tuned by Kazuki1450 from the Qwen3-1.7B-Base architecture. This model was trained using the GRPO method, which is designed to enhance mathematical reasoning capabilities. It is optimized for tasks requiring improved logical and mathematical processing, building upon the base Qwen3 model with a 32K context length.
Loading preview...