Kazuki1450/Qwen3-1.7B-Base_csum_3_10_tok_array_1p0_0p0_1p0_grpo_42_rule
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 18, 2026Architecture:Transformer Cold
Kazuki1450/Qwen3-1.7B-Base_csum_3_10_tok_array_1p0_0p0_1p0_grpo_42_rule is a 2 billion parameter language model fine-tuned from Qwen/Qwen3-1.7B-Base. This model was trained using the GRPO method, which is designed to enhance mathematical reasoning capabilities in language models. It is optimized for tasks requiring robust mathematical problem-solving and logical deduction, leveraging its 32768 token context length for complex inputs. This fine-tuned variant is suitable for applications demanding improved accuracy in quantitative and reasoning-based queries.
Loading preview...