hector-gr/RLCR-v4-ks-uniqueness-noece-noaurc-cold-math
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Mar 28, 2026Architecture:Transformer Cold
The hector-gr/RLCR-v4-ks-uniqueness-noece-noaurc-cold-math model is a 7.6 billion parameter language model fine-tuned from Qwen/Qwen2.5-7B, featuring a 32768 token context length. Developed by hector-gr, this model was trained using the TRL framework and incorporates the GRPO method, which is designed to enhance mathematical reasoning capabilities. Its primary strength lies in advanced reasoning tasks, particularly those requiring a deep understanding of mathematical concepts. This model is optimized for complex problem-solving and analytical applications.
Loading preview...