tlphams/zoyllm-7b-slimorca
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Dec 4, 2023License:cc-by-nc-sa-4.0Architecture:Transformer Open Weights Cold

ZoyLLM-7B-SlimOrca is a 7 billion parameter LoRA-finetuned generative text model developed by Pham Tung Lam and Nguyen Duc Nhan, built upon the Mistral-7B-v0.1 base architecture. It incorporates Grouped-Query Attention, Sliding-Window Attention, and a Byte-fallback BPE tokenizer, and is specifically fine-tuned using a chatml template for conversational AI. This model is designed for general text generation and chat applications, demonstrating performance that surpasses Llama 2 13B on tested benchmarks.

Loading preview...