juhwanlee/llmdo-Mistral-7B-case-5
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 7, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
The juhwanlee/llmdo-Mistral-7B-case-5 is a 7 billion parameter large language model developed by Juhwan Lee, fine-tuned from Mistral-7B-v0.1. This model is specifically optimized for data ordering tasks, leveraging architectural features like Grouped-Query Attention and Sliding-Window Attention. It was fine-tuned on a 100,000-sample subset of the Open-Orca dataset, making it suitable for specialized data arrangement applications.
Loading preview...