zzaen/day1-train-model
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Mar 25, 2026License:apache-2.0Architecture:Transformer Open Weights Warm
The zzaen/day1-train-model is a 0.5 billion parameter Qwen2-based instruction-tuned causal language model developed by zzaen. It was finetuned using Unsloth and Huggingface's TRL library, enabling 2x faster training. This model is optimized for efficient performance on tasks typically handled by smaller instruction-following models.
Loading preview...