BRlkl/distill-sft-qwen3-8b-full
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 27, 2026Architecture:Transformer Cold

BRlkl/distill-sft-qwen3-8b-full is an 8 billion parameter language model, fine-tuned from unsloth/Qwen3-8B using the TRL framework. This model is specifically trained with Supervised Fine-Tuning (SFT) to enhance its conversational and generative capabilities. With a 32K context length, it is designed for general text generation tasks, particularly excelling in interactive question-answering scenarios.

Loading preview...