mlabonne/Beagle14-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 15, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

mlabonne/Beagle14-7B is a 7 billion parameter language model created by mlabonne, formed by merging fblgit/UNA-TheBeagle-7b-v1 and argilla/distilabeled-Marcoro14-7B-slerp using a slerp merge method. This model demonstrates strong performance across various benchmarks, including AGIEval, GPT4All, and TruthfulQA, making it suitable for general-purpose reasoning and question-answering tasks. With a context length of 4096 tokens, it offers enhanced capabilities for processing longer inputs compared to many models in its class.

Loading preview...

Overview

mlabonne/Beagle14-7B is a 7-billion parameter language model developed by mlabonne, created through a strategic merge of two base models: fblgit/UNA-TheBeagle-7b-v1 and argilla/distilabeled-Marcoro14-7B-slerp. This merge was performed using the slerp (spherical linear interpolation) method, a technique often employed to combine the strengths of different models.

Key Capabilities & Performance

Beagle14-7B exhibits competitive performance across a suite of benchmarks, as evaluated using LLM AutoEval on the Nous suite and the Open LLM Leaderboard. It achieves an average score of 59.4 on the Nous suite, outperforming models like OpenHermes-2.5-Mistral-7B and NeuralHermes-2.5-Mistral-7B. Notable scores include:

  • AGIEval: 44.38
  • GPT4All: 76.53
  • TruthfulQA: 69.44
  • MMLU (5-Shot): 64.70
  • GSM8k (5-shot): 71.42

This model's evaluation results suggest strong reasoning, common sense, and factual recall abilities, positioning it as a robust option for various natural language processing tasks.

What Makes This Model Different?

Beagle14-7B stands out due to its origin as a carefully constructed merge, leveraging the distinct characteristics of its constituent models. Its performance metrics, particularly its higher scores on several benchmarks compared to other 7B and even some 10.7B models, indicate an effective combination of capabilities. The model's configuration details, including specific parameter weighting for self_attn and mlp layers during the merge, highlight a deliberate approach to optimizing its architecture.

Should You Use This Model?

This model is a strong candidate for use cases requiring a capable 7B model with good general reasoning and factual understanding. Its benchmark performance suggests it is well-suited for:

  • General-purpose chatbots and assistants
  • Question answering systems
  • Tasks requiring logical inference and common sense

For developers seeking a highly performant 7B model, especially one that excels in areas like TruthfulQA and GPT4All, Beagle14-7B offers a compelling option. For even stronger performance, users are encouraged to explore its DPO fine-tuned version, NeuralBeagle14-7B.