infinitylogesh/Qwen3-1.7B-GRPO-SRT-Math-12k-Stage-0
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Dec 14, 2025Architecture:Transformer Warm
The infinitylogesh/Qwen3-1.7B-GRPO-SRT-Math-12k-Stage-0 is a 2 billion parameter language model based on the Qwen3 architecture. This model is specifically designed and fine-tuned for mathematical reasoning and problem-solving tasks. Its primary differentiator lies in its specialized training for mathematical contexts, aiming for improved performance in numerical and logical operations. Developers can leverage this model for applications requiring robust mathematical capabilities.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–