juhwanlee/gemma-7B-alpaca-case-1-2
TEXT GENERATIONConcurrency Cost:1Model Size:8.5BQuant:FP8Ctx Length:8kPublished:Mar 25, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
The juhwanlee/gemma-7B-alpaca-case-1-2 is an 8.5 billion parameter large language model developed by Juhwan Lee. Based on the Mistral-7B-v0.1 architecture, it incorporates Grouped-Query Attention and Sliding-Window Attention. This model is specifically fine-tuned for data ordering tasks, utilizing a subset of the Open-Orca dataset. Its design focuses on efficient processing for structured data arrangement.
Loading preview...