JetBrains-Research/Qwen3-14B-am
TEXT GENERATIONConcurrency Cost:1Model Size:14BQuant:FP8Ctx Length:32kPublished:May 13, 2025Architecture:Transformer Cold

The JetBrains-Research/Qwen3-14B-am is a 14.8 billion parameter causal language model, developed by JetBrains Research, based on the Qwen3-14B architecture. This model uniquely incorporates an assistant mask token for improved identification of assistant-generated outputs, while retaining all original Qwen3 capabilities. It excels in reasoning, instruction-following, agent tasks, and multilingual support, with a native context length of 32,768 tokens, extendable to 131,072 tokens via YaRN scaling.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p