TwelfthStar/qwen3-8b-nothink-sft
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 31, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

Qwen3-8B is an 8.2 billion parameter causal language model developed by Qwen, featuring a unique capability to seamlessly switch between a 'thinking mode' for complex logical reasoning, math, and coding, and a 'non-thinking mode' for efficient general-purpose dialogue. This model significantly enhances reasoning, instruction-following, and agent capabilities, supporting over 100 languages. It is designed for optimal performance across diverse scenarios, from intricate problem-solving to natural conversational experiences, and natively supports a 32,768 token context length, extendable to 131,072 tokens with YaRN.

Loading preview...