Qwen3.5-397B-A17B is a 397 billion parameter multimodal causal language model developed by Qwen, featuring a unified vision-language foundation and an efficient hybrid architecture with Gated Delta Networks and sparse Mixture-of-Experts. It supports a native context length of 262,144 tokens, extensible up to 1,010,000 tokens using YaRN scaling. This model excels in multimodal reasoning, coding, agentic tasks, and visual understanding, with expanded support for 201 languages and dialects, making it suitable for global deployment and complex, long-horizon applications.
No reviews yet. Be the first to review!