jash404/qwen3-4b-half-subdivision-step90-clean
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026License:apache-2.0Architecture:Transformer Open Weights Loading

The Qwen3-4B model, developed by Qwen, is a 4.0 billion parameter causal language model with a native context length of 32,768 tokens, extendable to 131,072 tokens using YaRN. This model uniquely supports seamless switching between a 'thinking mode' for complex logical reasoning, math, and coding, and a 'non-thinking mode' for efficient general-purpose dialogue. It excels in reasoning capabilities, human preference alignment for creative writing and role-playing, and agentic tasks, supporting over 100 languages.

Loading preview...