hector-gr/RLCR-v4-ks-uniqueness-buf5k-noece-noaurc-cold-math
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Mar 28, 2026Architecture:Transformer Cold

The hector-gr/RLCR-v4-ks-uniqueness-buf5k-noece-noaurc-cold-math model is a 7.6 billion parameter language model, fine-tuned from Qwen/Qwen2.5-7B. Developed by hector-gr, it leverages the GRPO training method, known for enhancing mathematical reasoning in large language models. This model is specifically optimized for complex reasoning tasks, particularly those involving mathematical problem-solving. It offers a 32768-token context length, making it suitable for applications requiring deep analytical capabilities.

Loading preview...