allknowingroger/Qwenslerp4-14B
TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Nov 27, 2024Architecture:Transformer0.0K Cold

allknowingroger/Qwenslerp4-14B is a 14.8 billion parameter language model based on Qwen/Qwen2.5-14B, created by allknowingroger using the DARE TIES merge method. This model integrates several specialized Qwen2.5-14B variants to enhance performance across various reasoning and factual understanding tasks. It is specifically optimized for benchmarks like MATH, MUSR, GPQA, and IFEval, making it suitable for complex problem-solving and knowledge-intensive applications.

Loading preview...