princeton-nlp/Mistral-7B-Base-SFT-IPO
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:May 17, 2024Architecture:Transformer Cold
princeton-nlp/Mistral-7B-Base-SFT-IPO is a 7 billion parameter Mistral-based language model developed by princeton-nlp. This model is specifically fine-tuned using SimPO (Simple Preference Optimization with a Reference-Free Reward), a novel preference optimization technique. It is designed to demonstrate the effectiveness of SimPO as detailed in the associated research preprint, offering insights into advanced alignment methods.
Loading preview...