cjiao/goldengoose-method-v2-api-100

TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 28, 2026Architecture:Transformer Cold

The cjiao/goldengoose-method-v2-api-100 is a 1.5 billion parameter instruction-tuned language model, fine-tuned from Qwen/Qwen2.5-1.5B-Instruct. It utilizes the GRPO method, introduced in the DeepSeekMath paper, to enhance its capabilities. This model is optimized for tasks requiring advanced reasoning, particularly in mathematical contexts, and is suitable for applications needing precise and logical responses.

Loading preview...

Model Overview

cjiao/goldengoose-method-v2-api-100 is an instruction-tuned language model based on the Qwen2.5-1.5B-Instruct architecture, featuring 1.5 billion parameters and a 32K context length. It has been fine-tuned using the TRL framework.

Key Capabilities

  • Enhanced Reasoning: The model incorporates the GRPO (Gradient-based Reward Policy Optimization) method, a technique highlighted in the DeepSeekMath paper, which is designed to improve mathematical reasoning and problem-solving abilities.
  • Instruction Following: As an instruction-tuned model, it is adept at understanding and executing user prompts and instructions.
  • Efficient Performance: With 1.5 billion parameters, it offers a balance between performance and computational efficiency, making it suitable for various applications.

Training Details

The model's training procedure leveraged the TRL (Transformer Reinforcement Learning) framework, specifically version 0.19.1. The integration of the GRPO method suggests a focus on refining the model's logical and analytical output quality.

Good For

  • Mathematical Reasoning Tasks: Ideal for applications requiring logical deduction and mathematical problem-solving, benefiting from the GRPO fine-tuning.
  • Instruction-Based Generation: Suitable for generating responses based on explicit instructions, such as question answering or task completion.
  • Resource-Efficient Deployment: Its 1.5B parameter count makes it a viable option for scenarios where larger models might be too computationally intensive.