gguk2on/qwen2.5-7B-rlcr_g8_b384_math
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Apr 3, 2026Architecture:Transformer Cold

The gguk2on/qwen2.5-7B-rlcr_g8_b384_math model is a 7.6 billion parameter language model fine-tuned from Qwen/Qwen2.5-7B. It was trained using the TRL framework and incorporates the GRPO method, which is designed to enhance mathematical reasoning capabilities. This model is specifically optimized for complex mathematical problem-solving and reasoning tasks, leveraging techniques from DeepSeekMath. With a context length of 32768 tokens, it is suitable for applications requiring robust mathematical understanding and generation.

Loading preview...