akseljoonas/Qwen3-1.7B-SFT-s1K-lr0_0001
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Feb 27, 2026Architecture:Transformer Warm
The akseljoonas/Qwen3-1.7B-SFT-s1K-lr0_0001 is a 2 billion parameter language model, fine-tuned from the Qwen3-1.7B-Base architecture. Developed by akseljoonas, this model was trained using Supervised Fine-Tuning (SFT) on the simplescaling/s1K dataset. With a context length of 32768 tokens, it is designed for general text generation tasks, leveraging its fine-tuned capabilities for conversational responses.
Loading preview...