banghua/Qwen3-0.6B-SFT
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Apr 30, 2025Architecture:Transformer0.0K Warm

The banghua/Qwen3-0.6B-SFT model is a 0.8 billion parameter language model, likely a fine-tuned variant of the Qwen3 series, developed by banghua. With a substantial context length of 40960 tokens, it is designed for tasks requiring extensive contextual understanding. While specific differentiators are not detailed in the provided information, its large context window suggests potential optimization for long-form text processing, summarization, or complex question-answering over large documents.

Loading preview...