juhwanlee/llmdo-Mistral-7B-case-c-v1
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 4, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
The juhwanlee/llmdo-Mistral-7B-case-c-v1 is a 7 billion parameter large language model developed by Juhwan Lee, based on the Mistral-7B-v0.1 architecture. This model is specifically fine-tuned for data ordering tasks, utilizing a dataset sampled from Open-Orca. It incorporates architectural features like Grouped-Query Attention, Sliding-Window Attention, and a Byte-fallback BPE tokenizer, making it suitable for specialized data manipulation applications.
Loading preview...