princeton-nlp/Mistral-7B-Base-SFT-KTO
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:May 17, 2024Architecture:Transformer Cold

The princeton-nlp/Mistral-7B-Base-SFT-KTO is a 7 billion parameter language model based on the Mistral architecture, fine-tuned using the KTO (Kahneman-Tversky Optimization) method. Developed by princeton-nlp, this model is derived from research presented in the SimPO preprint, focusing on preference optimization with a reference-free reward. It is primarily designed for tasks benefiting from advanced alignment techniques, offering improved response quality and adherence to preferences.

Loading preview...