princeton-nlp/gemma-2-9b-it-SimPO
TEXT GENERATIONConcurrency Cost:1Model Size:9BQuant:FP8Ctx Length:16kPublished:Jul 16, 2024License:mitArchitecture:Transformer0.2K Open Weights Warm

The princeton-nlp/gemma-2-9b-it-SimPO model is a 9 billion parameter causal language model developed by Yu Meng, Mengzhou Xia, and Danqi Chen, fine-tuned from Google's Gemma-2-9B-IT. It utilizes the SimPO (Simple Preference Optimization) algorithm, an offline preference optimization method that aligns the reward function with generation likelihood without needing a reference model. This model is specifically designed to enhance performance on preference optimization datasets, making it suitable for tasks requiring nuanced response generation based on human preferences.

Loading preview...