chc2212/day1-train-model
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Apr 8, 2026License:apache-2.0Architecture:Transformer Open Weights Loading

The chc2212/day1-train-model is a 0.5 billion parameter Qwen2.5-based instruction-tuned causal language model, developed by chc2212. This model was finetuned using Unsloth and Huggingface's TRL library, enabling 2x faster training. It is designed for general language tasks, leveraging its efficient training methodology for rapid deployment.

Loading preview...