Fergus2000/wordle-grpo-Qwen3-1.7B
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Mar 24, 2026Architecture:Transformer Warm

The Fergus2000/wordle-grpo-Qwen3-1.7B is a 0.5 billion parameter language model fine-tuned from Qwen/Qwen2.5-0.5B-Instruct. It utilizes the GRPO training method, as introduced in the DeepSeekMath paper, to enhance mathematical reasoning capabilities. With a context length of 32768 tokens, this model is optimized for tasks requiring improved logical and mathematical processing. It is suitable for applications where robust reasoning on smaller models is beneficial.

Loading preview...