Baon2024/Qwen2.5-0.5B-SFT-training3
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Dec 28, 2025Architecture:Transformer Warm

Baon2024/Qwen2.5-0.5B-SFT-training3 is a 0.5 billion parameter language model fine-tuned from Qwen/Qwen2.5-0.5B. Developed by Baon2024, this model was trained using Supervised Fine-Tuning (SFT) on the HuggingFaceTB/smoltalk2 dataset. It is designed for general text generation tasks, leveraging its compact size and fine-tuned capabilities for efficient deployment. The model maintains a substantial context length of 131072 tokens, making it suitable for processing longer inputs.

Loading preview...