UCLA-AGI/Gemma-2-9B-It-SPPO-Iter2
TEXT GENERATIONConcurrency Cost:1Model Size:9BQuant:FP8Ctx Length:16kPublished:Jun 29, 2024License:gemmaArchitecture:Transformer0.0K Warm

The UCLA-AGI/Gemma-2-9B-It-SPPO-Iter2 is a 9 billion parameter GPT-like model, fine-tuned from google/gemma-2-9b-it using Self-Play Preference Optimization (SPPO) at its second iteration. This model leverages synthetic datasets derived from UltraFeedback prompts to enhance alignment. It is primarily English-language focused and designed for general instruction-following tasks, offering a 16384 token context length.

Loading preview...