akseljoonas/Qwen3-1.7B-SFT-s1K-lr1eneg05
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Feb 27, 2026Architecture:Transformer Warm

The akseljoonas/Qwen3-1.7B-SFT-s1K-lr1eneg05 model is a 1.7 billion parameter language model based on the Qwen3-1.7B-Base architecture. It has been fine-tuned using Supervised Fine-Tuning (SFT) on the simplescaling/s1K dataset, leveraging the TRL library. This model is designed for general text generation tasks, offering a compact yet capable solution for various natural language processing applications.

Loading preview...