SomayJalan/OpenRS-GRPO
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Nov 10, 2025Architecture:Transformer Cold

SomayJalan/OpenRS-GRPO is a 1.5 billion parameter language model fine-tuned from deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B with a 32768-token context length. It was trained using the GRPO method on the knoveleng/open-rs dataset, specializing in mathematical reasoning and complex problem-solving. This model is optimized for tasks requiring advanced logical deduction and numerical understanding.

Loading preview...