wandb/gemma-2b-zephyr-sft
TEXT GENERATIONConcurrency Cost:1Model Size:2.6BQuant:BF16Ctx Length:8kPublished:Feb 28, 2024License:gemma-terms-of-useArchitecture:Transformer0.0K Warm

wandb/gemma-2b-zephyr-sft is a 2.5 billion parameter GPT-like model fine-tuned by wandb, based on Google's Gemma 2B architecture. It applies the Zephyr Supervised Fine-Tuning (SFT) recipe to enhance its conversational and instruction-following capabilities. This model is primarily English-language focused and is suitable for general-purpose text generation and understanding tasks, demonstrating an average performance of 47.18 on the Open LLM Leaderboard.

Loading preview...