macadeliccc/MBX-7B-v3-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Jan 30, 2024License:ccArchitecture:Transformer0.0K Cold
macadeliccc/MBX-7B-v3-DPO is a 7 billion parameter causal language model, fine-tuned from flemmingmiguel/MBX-7B-v3 using Direct Preference Optimization (DPO). This model is optimized for conversational tasks and general instruction following, demonstrating improved performance over its base model on benchmarks like EQ-Bench and the Open LLM Leaderboard. With a context length of 8192 tokens, it is suitable for applications requiring nuanced responses and enhanced truthfulness.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p