jaygala24/Qwen3-1.7B-ReMax-math-reasoning
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Apr 13, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

jaygala24/Qwen3-1.7B-ReMax-math-reasoning is a 2 billion parameter language model, fine-tuned from Qwen3-1.7B, specifically optimized for mathematical reasoning tasks. It leverages the ReMax reinforcement learning algorithm without a KL penalty to enhance its ability to solve math problems. This model is designed for applications requiring step-by-step mathematical problem-solving capabilities.

Loading preview...