RJTPP/scot0402s-deepseek-1.5b-full
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 7, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
RJTPP/scot0402s-deepseek-1.5b-full is a 1.5 billion parameter Qwen2 model, finetuned by RJTPP. This model was trained using Unsloth and Huggingface's TRL library, enabling 2x faster training. It is designed for general language tasks, leveraging the DeepSeek-R1-Distill-Qwen-1.5B base architecture.
Loading preview...