gate369/Blurred-Beagle-7b-slerp
Blurred-Beagle-7b-slerp is a 7 billion parameter language model created by gate369, formed by merging alnrg2arg/blockchainlabs_7B_merged_test2_4 and 222gate/BrurryDog-7b-v0.1 using a slerp method. This model demonstrates a strong average performance of 74.80 on the Open LLM Leaderboard across various benchmarks, including reasoning, common sense, and mathematical tasks. With a 4096-token context length, it is suitable for general-purpose conversational AI and text generation where a balanced performance across diverse tasks is desired.
Loading preview...
Model Overview
Blurred-Beagle-7b-slerp is a 7 billion parameter language model developed by gate369. It was created by merging two base models, alnrg2arg/blockchainlabs_7B_merged_test2_4 and 222gate/BrurryDog-7b-v0.1, using a spherical linear interpolation (slerp) merge method. This approach combines the strengths of its constituent models to achieve a balanced performance profile.
Key Capabilities & Performance
This model has been evaluated on the Open LLM Leaderboard, achieving a notable average score of 74.80. Its performance highlights include:
- AI2 Reasoning Challenge (25-Shot): 72.78
- HellaSwag (10-Shot): 88.58
- MMLU (5-Shot): 64.95
- TruthfulQA (0-shot): 69.39
- Winogrande (5-shot): 83.19
- GSM8k (5-shot): 69.90
These scores indicate strong capabilities in reasoning, common sense understanding, truthfulness, and mathematical problem-solving, making it a versatile option for various NLP tasks.
Good For
- General-purpose text generation and conversational AI.
- Applications requiring balanced performance across diverse benchmarks.
- Tasks involving reasoning, common sense, and basic mathematical understanding.