s3nh/SeverusWestLake-7B-DPO
SeverusWestLake-7B-DPO is a 7 billion parameter language model created by s3nh, formed by merging FelixChao/Sectumsempra-7B-DPO and cognitivecomputations/WestLake-7B-v2-laser using the SLERP method. This model demonstrates strong general reasoning capabilities, achieving an average score of 75.42 on the Open LLM Leaderboard, with notable performance in tasks like HellaSwag (88.94) and Winogrande (86.11). It is designed for general-purpose applications requiring robust language understanding and generation.
Loading preview...
SeverusWestLake-7B-DPO Overview
SeverusWestLake-7B-DPO is a 7 billion parameter language model developed by s3nh. This model is a product of a sophisticated merge operation, combining two distinct pre-trained models: FelixChao/Sectumsempra-7B-DPO and cognitivecomputations/WestLake-7B-v2-laser. The merge was executed using the SLERP (Spherical Linear Interpolation) method, a technique known for effectively blending the strengths of constituent models.
Key Capabilities & Performance
This merged model exhibits strong performance across a range of benchmarks, as evaluated on the Open LLM Leaderboard. It achieved an average score of 75.42, indicating robust general-purpose language understanding and reasoning. Specific benchmark results include:
- AI2 Reasoning Challenge (25-Shot): 72.18
- HellaSwag (10-Shot): 88.94
- MMLU (5-Shot): 64.65
- TruthfulQA (0-shot): 71.49
- Winogrande (5-shot): 86.11
- GSM8k (5-shot): 69.14
These scores highlight its proficiency in common sense reasoning, factual recall, and problem-solving tasks.
Ideal Use Cases
SeverusWestLake-7B-DPO is well-suited for applications requiring a capable and efficient 7B parameter model. Its balanced performance across various benchmarks makes it a strong candidate for:
- General text generation and summarization
- Question answering and information extraction
- Reasoning-intensive tasks
- As a base for further fine-tuning on specific downstream applications