hector-gr/RLCR-v4-ks-uniqueness-hotpot-aliases-qwen35-balanced-fullnode-ga32
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Apr 8, 2026Architecture:Transformer Cold

hector-gr/RLCR-v4-ks-uniqueness-hotpot-aliases-qwen35-balanced-fullnode-ga32 is a 7.6 billion parameter language model fine-tuned from Qwen/Qwen2.5-7B. It was trained using the GRPO method, which is designed to enhance mathematical reasoning capabilities. This model is optimized for tasks requiring robust reasoning, particularly in mathematical contexts, leveraging its Qwen2.5 base and specialized training.

Loading preview...