juhwanlee/llmdo-Mistral-7B-case-6
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 5, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

The juhwanlee/llmdo-Mistral-7B-case-6 is a 7 billion parameter Large Language Model developed by Juhwan Lee, based on the Mistral-7B-v0.1 architecture. This model is specifically fine-tuned for data ordering tasks, utilizing Grouped-Query Attention and Sliding-Window Attention. It was fine-tuned on a random sample of 100,000 datasets from the Open-Orca dataset, making it specialized for data arrangement applications.

Loading preview...