Qwen/Qwen3.5-397B-A17B
TEXT GENERATIONConcurrency Cost:4Model Size:397BQuant:FP8Ctx Length:32kPublished:Feb 16, 2026License:apache-2.0 Vision Architecture:Transformer1.4K Open Weights Warm
Qwen3.5-397B-A17B is a multimodal causal language model developed by Qwen, featuring 397 billion total parameters with 17 billion activated. This model integrates a unified vision-language foundation and an efficient hybrid architecture, enabling cross-generational parity with Qwen3 and outperforming Qwen3-VL models across various benchmarks. It is designed for robust real-world adaptability, excelling in multimodal understanding, reasoning, coding, and agentic tasks with a native context length of 262,144 tokens, extensible up to 1,010,000 tokens.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–