wandb/gemma-2b-zephyr-dpo
TEXT GENERATIONConcurrency Cost:1Model Size:2.6BQuant:BF16Ctx Length:8kLicense:gemma-terms-of-useArchitecture:Transformer0.0K Warm

The wandb/gemma-2b-zephyr-dpo is a 2.6 billion parameter GPT-like language model developed by wandb, fine-tuned using the Zephyr DPO recipe. Primarily in English, this model is built upon a SFT finetuned Gemma 2B base. It is optimized for instruction following and conversational tasks through Direct Preference Optimization.

Loading preview...