gguk2on/qwen2.5-7B-rlvr_g8_b512
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Mar 23, 2026Architecture:Transformer Cold

The gguk2on/qwen2.5-7B-rlvr_g8_b512 is a 7.6 billion parameter language model, fine-tuned from Qwen/Qwen2.5-7B using the GRPO method. This model is specifically optimized for mathematical reasoning tasks, leveraging techniques introduced in the DeepSeekMath research. It is designed to enhance performance in complex problem-solving and logical deduction, making it suitable for applications requiring advanced analytical capabilities.

Loading preview...