puddledark/Qwen3-0.6B
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Mar 23, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

Qwen3-0.6B is a 0.6 billion parameter causal language model from the Qwen series, developed by Qwen. This model uniquely supports seamless switching between a 'thinking mode' for complex logical reasoning, math, and coding, and a 'non-thinking mode' for efficient general-purpose dialogue. It features a 32,768 token context length and excels in reasoning capabilities, human preference alignment, and agentic tasks across over 100 languages.

Loading preview...