Locutusque/Hyperion-3.0-Mistral-7B-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Mar 24, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Locutusque/Hyperion-3.0-Mistral-7B-DPO is a 7 billion parameter language model based on Mistral-7B-v0.1, fine-tuned using Direct Preference Optimization (DPO) on 20,000 GPT-4 generated preference pairs. It is designed for superior performance across complex tasks including question answering, conversational AI, code generation, medical text comprehension, mathematical reasoning, and logical reasoning. The model offers an 8192 token context length and achieves an MMLU score of 0.5833, demonstrating broad multi-domain proficiency.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p