akseljoonas/Qwen3-1.7B-SFT-s1K-lr2eneg05
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Feb 27, 2026Architecture:Transformer Warm

The akseljoonas/Qwen3-1.7B-SFT-s1K-lr2eneg05 is a 1.7 billion parameter causal language model, fine-tuned from Qwen/Qwen3-1.7B-Base. This model was specifically trained using Supervised Fine-Tuning (SFT) on the simplescaling/s1K dataset, leveraging the TRL framework. It is designed for general text generation tasks, offering a compact yet capable solution for various natural language processing applications.

Loading preview...