kaist-ai/mistral-orpo-beta
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Mar 12, 2024License:mitArchitecture:Transformer0.0K Open Weights Cold

kaist-ai/mistral-orpo-beta is a 7 billion parameter language model developed by KAIST AI, fine-tuned from Mistral-7B-v0.1 using the Odds Ratio Preference Optimization (ORPO) method. This model directly learns preferences without a supervised fine-tuning warmup, distinguishing it from traditional alignment techniques. It is specifically optimized for conversational AI and instruction following, demonstrating strong performance on benchmarks like MT-Bench and AlpacaEval.

Loading preview...