razy101/qwen3-0.6b-gpt4-distilled
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The razy101/qwen3-0.6b-gpt4-distilled is a 0.8 billion parameter Qwen3 model, fine-tuned by razy101. This model was trained 2x faster using Unsloth and Huggingface's TRL library, offering a highly efficient and optimized small language model. It is designed for tasks requiring a compact yet capable model, leveraging its efficient training methodology.

Loading preview...