macadeliccc/MBX-7B-v3-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Jan 30, 2024License:ccArchitecture:Transformer0.0K Cold

macadeliccc/MBX-7B-v3-DPO is a 7 billion parameter causal language model, fine-tuned from flemmingmiguel/MBX-7B-v3 using Direct Preference Optimization (DPO). This model is optimized for conversational tasks and general instruction following, demonstrating improved performance over its base model on benchmarks like EQ-Bench and the Open LLM Leaderboard. With a context length of 8192 tokens, it is suitable for applications requiring nuanced responses and enhanced truthfulness.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p