AlazarM/trenches-us-qwen3-8b-real
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 8, 2026Architecture:Transformer Cold
AlazarM/trenches-us-qwen3-8b-real is an 8 billion parameter language model fine-tuned from Qwen/Qwen3-8B, utilizing the TRL framework. This model was specifically trained with the GRPO method, as introduced in the DeepSeekMath paper, to enhance its mathematical reasoning capabilities. With a context length of 32768 tokens, it is optimized for tasks requiring advanced mathematical problem-solving and logical deduction.
Loading preview...