The JetBrains-Research/Qwen3-14B-am is a 14.8 billion parameter causal language model, developed by JetBrains Research, based on the Qwen3-14B architecture. This model uniquely incorporates an assistant mask token for improved identification of assistant-generated outputs, while retaining all original Qwen3 capabilities. It excels in reasoning, instruction-following, agent tasks, and multilingual support, with a native context length of 32,768 tokens, extendable to 131,072 tokens via YaRN scaling.
No reviews yet. Be the first to review!