mlabonne/AlphaMonarch-7B
mlabonne/AlphaMonarch-7B is a 7 billion parameter DPO fine-tuned language model developed by mlabonne, based on a merge of several models including NeuralMonarch-7B. It features an 8k context window and is optimized to retain strong reasoning abilities while significantly improving conversational capabilities. This model excels in instruction following, reasoning, and conversational tasks, making it suitable for general-purpose chat, roleplay, and storytelling applications.
Loading preview...
AlphaMonarch-7B: Enhanced Conversational and Reasoning 7B Model
AlphaMonarch-7B is a 7 billion parameter language model developed by mlabonne, created through a DPO (Direct Preference Optimization) fine-tuning of mlabonne/NeuralMonarch-7B. This model is a merge of several base models, including OmniTruthyBeagle-7B-v0, NeuBeagle-7B, and NeuralOmniBeagle-7B, utilizing the argilla/OpenHermes2.5-dpo-binarized-alpha preference dataset.
Key Capabilities & Differentiators
- Balanced Performance: Designed to offer a strong balance between reasoning abilities and conversational fluency, addressing a common trade-off in 7B models.
- Instruction Following: Demonstrates high proficiency in following instructions, making it versatile for various tasks.
- Context Window: Supports an 8k token context window, suitable for longer interactions.
- Top-tier 7B Performance: Achieves leading scores among 7B models on benchmarks like Nous's benchmark suite (62.74 average), EQ-bench (outperforming some 70B/120B models), and competitive MT-Bench scores.
Use Cases
AlphaMonarch-7B is well-suited for applications requiring:
- Conversational AI: Engaging in natural and coherent dialogues.
- Roleplay and Storytelling: Generating creative and consistent narratives.
- Reasoning Tasks: Handling complex queries that require logical deduction.
- General-purpose Chatbots: Providing informative and context-aware responses.
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.