mlabonne/Marcoro14-7B-slerp

Warm
Public
7B
FP8
4096
Dec 29, 2023
License: cc-by-nc-4.0
Hugging Face
Overview

Marcoro14-7B-slerp: A High-Performing 7B LLM

Marcoro14-7B-slerp is a 7 billion parameter language model created by mlabonne through a merge of two base models: AIDC-ai-business/Marcoroni-7B-v3 and EmbeddedLLM/Mistral-7B-Merge-14-v0.1. This merge was performed using the slerp method, a technique for interpolating between model weights.

Key Capabilities & Performance

This model stands out for its strong benchmark performance, positioning it as a top-tier 7B model on the Open LLM Leaderboard. It demonstrates significant improvements over models like OpenHermes-2.5-Mistral-7B across several evaluation suites:

  • Open LLM Leaderboard: Achieves an average score of 73.01, with notable results in AI2 Reasoning Challenge (69.80), HellaSwag (87.13), MMLU (65.11), and GSM8k (70.89).
  • Nous' Benchmark Suite: Scores an average of 57.67, outperforming OpenHermes-2.5-Mistral-7B by over 5 points. Specific gains include +1.59 on AGIEval, +3.12 on GPT4ALL, +11.11 on TruthfulQA, and +4.68 on Bigbench.
  • Reasoning: Shows strong capabilities in reasoning tasks, as indicated by its AGIEval and Bigbench scores, including logical deduction and causal judgment.
  • Truthfulness: Achieves a TruthfulQA mc2 score of 64.15, suggesting a good ability to generate factually correct responses.

Good For

  • Applications requiring a highly capable 7B model for general-purpose text generation and understanding.
  • Tasks that benefit from strong reasoning and analytical skills.
  • Developers looking for a performant and efficient model in the 7B parameter class.