Qwen/Qwen2.5-14B
TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Sep 15, 2024License:apache-2.0Architecture:Transformer0.1KOpen Weights Warm

Qwen/Qwen2.5-14B is a 14.7 billion parameter causal language model developed by Qwen, featuring a 131,072 token context length. This base model, part of the Qwen2.5 series, significantly improves upon Qwen2 with enhanced knowledge, coding, and mathematics capabilities, alongside better instruction following and long-text generation. It is designed for pretraining and further fine-tuning, offering multilingual support for over 29 languages.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p