cloudyu/google-gemma-7b-it-dpo-v1
TEXT GENERATIONConcurrency Cost:1Model Size:8.5BQuant:FP8Ctx Length:8kPublished:Feb 23, 2024License:gemma-terms-of-useArchitecture:Transformer Cold

The cloudyu/google-gemma-7b-it-dpo-v1 is an 8.5 billion parameter language model, fine-tuned using Direct Preference Optimization (DPO) on the google/gemma-7b-it base model. This DPO fine-tuning process, utilizing the jondurbin/truthy-dpo-v0.1 dataset, aims to align the model's outputs more closely with human preferences. It is designed for general-purpose conversational AI and instruction-following tasks, leveraging its 8192-token context length for coherent and extended interactions.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p