AALF/gemma-2-27b-it-SimPO-37K
TEXT GENERATIONConcurrency Cost:2Model Size:27BQuant:FP8Ctx Length:32kPublished:Aug 13, 2024License:gemmaArchitecture:Transformer0.0K Warm

AALF/gemma-2-27b-it-SimPO-37K is a fine-tuned version of Google's Gemma 2 27B instruction-tuned model. This model was enhanced using the SimPO framework with on-policy preference data generated from the HuggingFaceH4/ultrafeedback_binarized dataset. It is optimized for improved response quality based on reward model feedback, making it suitable for conversational AI and instruction-following tasks.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p