zzmini/day1-train-model
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Apr 8, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The zzmini/day1-train-model is a 0.5 billion parameter Qwen2.5-Instruct model, developed by zzmini, fine-tuned for instruction following. It was trained using Unsloth and Huggingface's TRL library, achieving 2x faster training speeds. This model is optimized for efficient deployment in scenarios requiring a compact yet capable instruction-tuned language model.
Loading preview...