Qwen/Qwen2.5-72B-Instruct
TEXT GENERATIONConcurrency Cost:4Model Size:72.7BQuant:FP8Ctx Length:32kPublished:Sep 16, 2024License:qwenArchitecture:Transformer0.9K Warm

Qwen2.5-72B-Instruct is a 72.7 billion parameter instruction-tuned causal language model developed by Qwen, built upon the Qwen2 architecture. It features significant improvements in knowledge, coding, and mathematics, alongside enhanced instruction following and long text generation up to 8K tokens. The model supports a 131,072-token context length and offers robust multilingual capabilities across 29 languages, making it suitable for complex, diverse language tasks.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p