KaraKaraWitch/SteyrCannon-0.2-Qwen2.5-72b
TEXT GENERATIONConcurrency Cost:4Model Size:72.7BQuant:FP8Ctx Length:32kPublished:Nov 23, 2024Architecture:Transformer0.0K Warm
KaraKaraWitch/SteyrCannon-0.2-Qwen2.5-72b is a 72.7 billion parameter language model merge, built upon the Qwen2.5 architecture with a 131072 token context length. Created using the TIES merge method, it combines anthracite-org/magnum-v4-72b and EVA-UNIT-01/EVA-Qwen2.5-72B-v0.2, using the latter as its base. This model is designed for general language tasks, leveraging the strengths of its merged components.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
top_p
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p