unsloth/Qwen3-0.6B

Cold
Public
0.8B
BF16
40960
Hugging Face
Overview

Qwen3-0.6B Overview

Qwen3-0.6B is a 0.6 billion parameter causal language model from the Qwen3 series, developed by Qwen. It introduces a novel feature allowing seamless switching between a 'thinking mode' for complex logical reasoning, math, and coding, and a 'non-thinking mode' for efficient, general-purpose dialogue. This dual-mode functionality ensures optimal performance across diverse scenarios.

Key Capabilities

  • Enhanced Reasoning: Significantly improved performance in mathematics, code generation, and commonsense logical reasoning compared to previous Qwen models.
  • Human Preference Alignment: Excels in creative writing, role-playing, multi-turn dialogues, and instruction following, providing a natural and engaging conversational experience.
  • Agent Capabilities: Demonstrates strong tool-calling abilities, enabling precise integration with external tools in both thinking and non-thinking modes, achieving leading performance in complex agent-based tasks among open-source models.
  • Multilingual Support: Supports over 100 languages and dialects with robust capabilities for multilingual instruction following and translation.

Usage and Best Practices

The model supports a context length of 32,768 tokens. Users can explicitly enable or disable the thinking mode via enable_thinking in the tokenizer's chat template, or dynamically switch modes within user prompts using /think and /no_think tags. For optimal performance, specific sampling parameters are recommended for each mode (e.g., Temperature=0.6 for thinking mode, Temperature=0.7 for non-thinking mode). The model is compatible with Hugging Face transformers (version 4.51.0 or newer) and can be deployed with vLLM or sglang for OpenAI-compatible API endpoints.