ale-bay/zephyr-2b-gemma-sft
TEXT GENERATIONConcurrency Cost:1Model Size:2.6BQuant:BF16Ctx Length:8kLicense:gemmaArchitecture:Transformer Warm

ale-bay/zephyr-2b-gemma-sft is a 2.6 billion parameter language model fine-tuned from google/gemma-2b. This model was instruction-tuned using the HuggingFaceH4/deita-10k-v0-sft dataset. It is designed for general language generation tasks, demonstrating a validation loss of 1.0529 after 3 epochs of training.

Loading preview...