rubenroy/Zurich-14B-GCv2-5m
TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Jan 31, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm
rubenroy/Zurich-14B-GCv2-5m is a 14.7 billion parameter causal language model, fine-tuned by Ruben Roy from Alibaba's Qwen 2.5 14B Instruct base model. It utilizes a Transformer architecture with RoPE, SwiGLU, RMSNorm, and Attention QKV bias. This model is specifically fine-tuned on the GammaCorpus v2-5m dataset, which consists of structured and filtered multi-turn conversations, aiming to outperform similarly sized models in conversational tasks.
Loading preview...