juhwanlee/gemma-7B-alpaca-case-3-2
TEXT GENERATIONConcurrency Cost:1Model Size:8.5BQuant:FP8Ctx Length:8kPublished:Mar 25, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
The juhwanlee/gemma-7B-alpaca-case-3-2 is an 8.5 billion parameter large language model developed by Juhwan Lee, based on the Gemma-7B architecture. This model has been fine-tuned specifically for data ordering tasks, utilizing a randomly sampled subset of the Open-Orca dataset. It incorporates architectural features such as Grouped-Query Attention, Sliding-Window Attention, and a byte-fallback BPE tokenizer, making it suitable for specialized data processing applications.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–