GreatGoose/Qwen2.5-0.5B-Instruct-distill-3epoch

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Jan 5, 2026Architecture:Transformer Warm

GreatGoose/Qwen2.5-0.5B-Instruct-distill-3epoch is a 0.5 billion parameter instruction-tuned causal language model, distilled from Qwen/Qwen2.5-0.5B-Instruct. Developed by GreatGoose, this model leverages TRL for training and incorporates GOLD for on-policy distillation. It is designed for general text generation tasks, offering a compact yet capable solution for various applications.

Loading preview...

Model Overview

GreatGoose/Qwen2.5-0.5B-Instruct-distill-3epoch is a compact 0.5 billion parameter instruction-tuned language model. It is a distilled version of the Qwen/Qwen2.5-0.5B-Instruct base model, fine-tuned using the TRL (Transformer Reinforcement Learning) library. This distillation process, specifically utilizing GOLD (General On-Policy Logit Distillation), aims to transfer knowledge from a larger or more capable model into this smaller, more efficient variant.

Key Capabilities

  • Instruction Following: Designed to respond to user instructions effectively, making it suitable for conversational agents and task-oriented applications.
  • Text Generation: Capable of generating coherent and contextually relevant text based on given prompts.
  • Efficient Deployment: As a 0.5B parameter model, it offers a balance between performance and computational efficiency, making it suitable for environments with limited resources.

Training Details

The model was trained with GOLD, a method for on-policy distillation, and utilized the TRL framework (version 0.26.2). The training process involved specific versions of key libraries including Transformers (4.57.3), Pytorch (2.9.1), Datasets (4.4.2), and Tokenizers (0.22.1).

Good For

  • Applications requiring a lightweight, instruction-following language model.
  • Scenarios where computational resources are constrained, but reasonable text generation capabilities are needed.
  • Experimentation with distilled models for various NLP tasks.