hydra-project/ChatHercules-2.5-Mistral-7B
ChatHercules-2.5-Mistral-7B is a 7 billion parameter language model created by hydra-project, formed by merging Locutusque/Hercules-2.5-Mistral-7B and openchat/openchat-3.5-0106. This model leverages a slerp merge method to combine the strengths of its base models, achieving an average performance of 68.24 on the Open LLM Leaderboard. It is designed for general conversational AI tasks, demonstrating balanced capabilities across reasoning, common sense, and language understanding benchmarks.
Loading preview...
ChatHercules-2.5-Mistral-7B Overview
ChatHercules-2.5-Mistral-7B is a 7 billion parameter language model developed by hydra-project. It is a product of merging two distinct models: Locutusque/Hercules-2.5-Mistral-7B and openchat/openchat-3.5-0106, utilizing a slerp merge method via LazyMergekit. This approach aims to combine the beneficial characteristics of both foundational models into a single, cohesive unit.
Key Capabilities & Performance
This model demonstrates solid performance across a range of benchmarks, as evaluated on the Open LLM Leaderboard. It achieves an average score of 68.24, indicating balanced capabilities. Notable scores include:
- AI2 Reasoning Challenge (25-Shot): 65.10
- HellaSwag (10-Shot): 84.61
- MMLU (5-Shot): 65.35
- Winogrande (5-Shot): 81.85
- GSM8k (5-Shot): 64.97
These results suggest proficiency in common sense reasoning, language understanding, and mathematical problem-solving, making it a versatile option for various applications.
Good For
- General-purpose conversational AI: Its balanced performance makes it suitable for chatbots and interactive agents.
- Reasoning tasks: Scores on ARC and GSM8k indicate good logical and mathematical reasoning abilities.
- Language understanding and generation: Strong HellaSwag and Winogrande scores point to robust capabilities in these areas.