juhwanlee/experiment2-cause-v1
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 5, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

The juhwanlee/experiment2-cause-v1 is a 7 billion parameter large language model developed by Juhwan Lee. It is based on the Mistral-7B-v0.1 architecture, featuring Grouped-Query Attention and Sliding-Window Attention. This model is specifically fine-tuned for data ordering tasks, utilizing a randomly sampled subset of the Open-Orca dataset for its training.

Loading preview...