JetBrains-Research/Qwen3-32B-am
TEXT GENERATIONConcurrency Cost:2Model Size:32BQuant:FP8Ctx Length:32kPublished:May 14, 2025Architecture:Transformer Cold

JetBrains-Research/Qwen3-32B-am is a 32.8 billion parameter causal language model developed by JetBrains Research, based on the Qwen3 architecture. This model uniquely integrates an assistant mask token for improved identification of assistant-generated outputs, while preserving the original Qwen3's advanced reasoning, instruction-following, and agent capabilities. It supports seamless switching between a 'thinking mode' for complex tasks like math and coding, and a 'non-thinking mode' for efficient general dialogue, with a native context length of 32,768 tokens.

Loading preview...