shadowml/BeagSake-7B

Cold
Public
7B
FP8
4096
License: cc-by-nc-4.0
Hugging Face
Overview

Model Overview

shadowml/BeagSake-7B is a 7 billion parameter language model developed by shadowml. It is a merged model, combining the strengths of shadowml/BeagleSempra-7B and shadowml/WestBeagle-7B using the slerp (spherical linear interpolation) merge method via LazyMergekit. This approach allows for a blend of the characteristics of its constituent models.

Key Capabilities

BeagSake-7B demonstrates solid performance across various benchmarks, indicating strong general language understanding and reasoning. Key evaluation results from the Open LLM Leaderboard include:

  • Average Score: 75.38
  • AI2 Reasoning Challenge (25-Shot): 72.44
  • HellaSwag (10-Shot): 88.39
  • MMLU (5-Shot): 65.23
  • TruthfulQA (0-shot): 72.27
  • Winogrande (5-shot): 82.16
  • GSM8k (5-shot): 71.80

These scores suggest proficiency in common sense reasoning, factual recall, and mathematical problem-solving.

Good For

  • General-purpose text generation: Creating coherent and contextually relevant text.
  • Question Answering: Responding to queries based on its training data.
  • Reasoning tasks: Handling tasks that require logical inference and problem-solving, as indicated by its AI2 Reasoning Challenge and GSM8k scores.