cjiao/goldengoose-corr-v4-1.00-200

TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:May 2, 2026Architecture:Transformer Cold

The cjiao/goldengoose-corr-v4-1.00-200 model is a 1.5 billion parameter instruction-tuned language model, fine-tuned from Qwen/Qwen2.5-1.5B-Instruct. It was trained using the GRPO method, which is designed to enhance mathematical reasoning capabilities. With a context length of 32768 tokens, this model is optimized for tasks requiring robust mathematical and logical problem-solving.

Loading preview...

Model Overview

cjiao/goldengoose-corr-v4-1.00-200 is a 1.5 billion parameter language model, fine-tuned from the Qwen/Qwen2.5-1.5B-Instruct architecture. This model leverages a substantial 32768-token context window, making it suitable for processing longer inputs and maintaining coherence over extended interactions.

Key Capabilities and Training

The primary differentiator of this model lies in its training methodology. It was fine-tuned using GRPO (Gradient-based Reward Policy Optimization), a technique introduced in the paper "DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models." This training approach specifically aims to improve the model's proficiency in mathematical reasoning and complex problem-solving.

When to Use This Model

  • Mathematical Reasoning Tasks: Ideal for applications requiring strong logical deduction and mathematical problem-solving, benefiting from the GRPO training.
  • Instruction Following: As an instruction-tuned model, it is designed to accurately follow user prompts and generate relevant responses.
  • Long Context Applications: Its 32768-token context length supports tasks that involve processing or generating extensive text, such as summarization of long documents or detailed conversational agents.