gate369/Blurred-Beagle-7b-slerp

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 20, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Blurred-Beagle-7b-slerp is a 7 billion parameter language model created by gate369, formed by merging alnrg2arg/blockchainlabs_7B_merged_test2_4 and 222gate/BrurryDog-7b-v0.1 using a slerp method. This model demonstrates a strong average performance of 74.80 on the Open LLM Leaderboard across various benchmarks, including reasoning, common sense, and mathematical tasks. With a 4096-token context length, it is suitable for general-purpose conversational AI and text generation where a balanced performance across diverse tasks is desired.

Loading preview...

Model Overview

Blurred-Beagle-7b-slerp is a 7 billion parameter language model developed by gate369. It was created by merging two base models, alnrg2arg/blockchainlabs_7B_merged_test2_4 and 222gate/BrurryDog-7b-v0.1, using a spherical linear interpolation (slerp) merge method. This approach combines the strengths of its constituent models to achieve a balanced performance profile.

Key Capabilities & Performance

This model has been evaluated on the Open LLM Leaderboard, achieving a notable average score of 74.80. Its performance highlights include:

  • AI2 Reasoning Challenge (25-Shot): 72.78
  • HellaSwag (10-Shot): 88.58
  • MMLU (5-Shot): 64.95
  • TruthfulQA (0-shot): 69.39
  • Winogrande (5-shot): 83.19
  • GSM8k (5-shot): 69.90

These scores indicate strong capabilities in reasoning, common sense understanding, truthfulness, and mathematical problem-solving, making it a versatile option for various NLP tasks.

Good For

  • General-purpose text generation and conversational AI.
  • Applications requiring balanced performance across diverse benchmarks.
  • Tasks involving reasoning, common sense, and basic mathematical understanding.