Gille/StrangeMerges_28-7B-dare_ties
Gille/StrangeMerges_28-7B-dare_ties is a 7 billion parameter language model created by Gille, developed using the dare_ties merge method on a CultriX/MonaTrix-v4 base model. It integrates components from eren23/ogno-monarch-jaskier-merge-7b-v2 and Gille/StrangeMerges_25-7B-dare_ties. This model achieves an average score of 75.86 on the Open LLM Leaderboard, demonstrating strong general reasoning and language understanding capabilities, making it suitable for a variety of common NLP tasks.
Loading preview...
StrangeMerges_28-7B-dare_ties Overview
StrangeMerges_28-7B-dare_ties is a 7 billion parameter language model developed by Gille. This model was created using the dare_ties merge method via LazyMergekit, combining several existing models to enhance performance. Its base model is CultriX/MonaTrix-v4, and it integrates specific contributions from eren23/ogno-monarch-jaskier-merge-7b-v2 and Gille/StrangeMerges_25-7B-dare_ties.
Key Capabilities
- General-purpose language understanding: Designed for a broad range of NLP tasks.
- Reasoning: Achieves a score of 72.18 on the AI2 Reasoning Challenge (25-Shot).
- Common sense reasoning: Scores 89.08 on HellaSwag (10-Shot) and 83.50 on Winogrande (5-shot).
- Mathematical problem-solving: Demonstrates capability with a GSM8k (5-shot) score of 68.16.
- Instruction following: Supports standard chat template usage for text generation.
Performance Highlights
This model has been evaluated on the Open LLM Leaderboard, achieving an average score of 75.86. Notable individual benchmark results include:
- MMLU (5-Shot): 64.68
- TruthfulQA (0-shot): 77.55
Good for
- Applications requiring robust general language understanding and generation.
- Tasks benefiting from strong reasoning and common sense capabilities.
- Developers looking for a merged model with competitive performance in the 7B parameter class.