gagan3012/MetaModel

Warm
Public
10.7B
FP8
4096
License: apache-2.0
Hugging Face
Overview

MetaModel Overview

MetaModel is a 10.7 billion parameter language model developed by gagan3012. It is a product of merging two distinct models, jeonsworld/CarbonVillain-en-10.7B-v4 and kekmodel/StopCarbon-10.7B-v5, utilizing the slerp merge method via mergekit. This merging strategy aims to combine the strengths of its constituent models.

Key Capabilities & Performance

Evaluated on the Open LLM Leaderboard, MetaModel achieved an average score of 74.4. Notable benchmark results include:

  • ARC (25-shot): 71.08
  • HellaSwag (10-shot): 88.45
  • MMLU (5-shot): 66.26
  • TruthfulQA (0-shot): 71.84
  • Winogrande (5-shot): 83.43
  • GSM8K (5-shot): 65.35

These scores indicate a strong general understanding and reasoning capability across a variety of tasks, from common sense reasoning to academic subjects and mathematical problem-solving.

Use Cases

MetaModel is well-suited for applications requiring a versatile language model with solid performance across a broad spectrum of tasks. Its balanced benchmark results suggest it can be effectively used for:

  • General text generation and comprehension
  • Question answering
  • Reasoning tasks
  • Educational support in various subjects