Qwen/Qwen3.5-397B-A17B
TEXT GENERATIONConcurrency Cost:4Model Size:397BQuant:FP8Ctx Length:32kPublished:Feb 16, 2026License:apache-2.0 Vision Architecture:Transformer1.4K Open Weights Warm

Qwen3.5-397B-A17B is a multimodal causal language model developed by Qwen, featuring 397 billion total parameters with 17 billion activated. This model integrates a unified vision-language foundation and an efficient hybrid architecture, enabling cross-generational parity with Qwen3 and outperforming Qwen3-VL models across various benchmarks. It is designed for robust real-world adaptability, excelling in multimodal understanding, reasoning, coding, and agentic tasks with a native context length of 262,144 tokens, extensible up to 1,010,000 tokens.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p