Qwen/Qwen3-Next-80B-A3B-Instruct
TEXT GENERATIONConcurrency Cost:4Model Size:80BQuant:FP8Ctx Length:32kPublished:Sep 9, 2025License:apache-2.0Architecture:Transformer1.0K Open Weights Warm

Qwen/Qwen3-Next-80B-A3B-Instruct is an 80 billion parameter instruction-tuned causal language model developed by Qwen, featuring a hybrid attention mechanism and high-sparsity Mixture-of-Experts (MoE) architecture. It is designed for efficient context modeling and ultra-long context lengths up to 262,144 tokens natively, with extensibility to 1 million tokens via YaRN. This model excels in parameter efficiency and inference speed, particularly for long-context tasks, and demonstrates strong performance across knowledge, reasoning, coding, and alignment benchmarks.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p