ryzax/xxx
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Jan 21, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

Qwen3-0.6B is a 0.6 billion parameter causal language model developed by Qwen, featuring a 32,768 token context length. This model uniquely supports seamless switching between a 'thinking mode' for complex logical reasoning, math, and coding, and a 'non-thinking mode' for efficient general-purpose dialogue. It demonstrates enhanced reasoning capabilities, superior human preference alignment for creative writing and role-playing, and strong agent capabilities with external tool integration. Qwen3-0.6B also supports over 100 languages and dialects for multilingual instruction following and translation.

Loading preview...