juhwanlee/gemma-7B-alpaca-case-2-2

TEXT GENERATIONConcurrency Cost:1Model Size:8.5BQuant:FP8Ctx Length:8kPublished:Mar 25, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

The juhwanlee/gemma-7B-alpaca-case-2-2 is an 8.5 billion parameter large language model developed by Juhwan Lee. Based on the Gemma-7B architecture, it features Grouped-Query Attention, Sliding-Window Attention, and a Byte-fallback BPE tokenizer. This model is specifically fine-tuned for data ordering tasks, utilizing a randomly sampled subset of the Open-Orca dataset. Its primary application is for testing data ordering processes.

Loading preview...

Model Overview

The juhwanlee/gemma-7B-alpaca-case-2-2 is an 8.5 billion parameter Large Language Model developed by Juhwan Lee. It is built upon the Gemma-7B transformer architecture, incorporating features such as Grouped-Query Attention, Sliding-Window Attention, and a Byte-fallback BPE tokenizer. This model has undergone fine-tuning specifically for data ordering tasks.

Key Capabilities

  • Data Ordering: The model's primary specialization is in handling and processing data ordering tasks, having been fine-tuned on a subset of the Open-Orca dataset for this purpose.
  • Gemma-7B Foundation: Leverages the robust architecture of Gemma-7B, providing a strong base for language understanding and generation.

Training Details

The model was fine-tuned using a random sample of 100,000 data points from the Open-Orca dataset, focusing on optimizing its performance for data ordering.

Licensing

This model is released under the Apache License 2.0, allowing for broad use and distribution.