mlabonne/AlphaMonarch-7B

Loading
Public
7B
FP8
8192
Feb 14, 2024
License: cc-by-nc-4.0
Hugging Face
Overview

AlphaMonarch-7B: Enhanced Conversational and Reasoning 7B Model

AlphaMonarch-7B is a 7 billion parameter language model developed by mlabonne, created through a DPO (Direct Preference Optimization) fine-tuning of mlabonne/NeuralMonarch-7B. This model is a merge of several base models, including OmniTruthyBeagle-7B-v0, NeuBeagle-7B, and NeuralOmniBeagle-7B, utilizing the argilla/OpenHermes2.5-dpo-binarized-alpha preference dataset.

Key Capabilities & Differentiators

  • Balanced Performance: Designed to offer a strong balance between reasoning abilities and conversational fluency, addressing a common trade-off in 7B models.
  • Instruction Following: Demonstrates high proficiency in following instructions, making it versatile for various tasks.
  • Context Window: Supports an 8k token context window, suitable for longer interactions.
  • Top-tier 7B Performance: Achieves leading scores among 7B models on benchmarks like Nous's benchmark suite (62.74 average), EQ-bench (outperforming some 70B/120B models), and competitive MT-Bench scores.

Use Cases

AlphaMonarch-7B is well-suited for applications requiring:

  • Conversational AI: Engaging in natural and coherent dialogues.
  • Roleplay and Storytelling: Generating creative and consistent narratives.
  • Reasoning Tasks: Handling complex queries that require logical deduction.
  • General-purpose Chatbots: Providing informative and context-aware responses.