Model Overview
mlabonne/NeuralPipe-7B-slerp is a 7 billion parameter language model developed by mlabonne, created through a merge of two base models: OpenPipe/mistral-ft-optimized-1218 and mlabonne/NeuralHermes-2.5-Mistral-7B. This merge was performed using the slerp (spherical linear interpolation) method, which combines the strengths of its constituent models to enhance overall performance.
Key Capabilities
- General Language Understanding: The model exhibits robust capabilities across various NLP tasks, as indicated by its performance on the Open LLM Leaderboard.
- Reasoning: Achieves a score of 67.75 on the AI2 Reasoning Challenge (25-Shot) and 69.75 on GSM8k (5-shot), suggesting proficiency in logical and mathematical reasoning.
- Context Handling: Supports a context length of 4096 tokens, allowing it to process and generate longer sequences of text.
- Instruction Following: Inherits instruction-following capabilities from its base models, making it suitable for chat and prompt-based interactions.
Performance Highlights
On the Hugging Face Open LLM Leaderboard, NeuralPipe-7B-slerp achieved an average score of 71.17. Specific benchmark results include:
- HellaSwag (10-Shot): 86.15
- MMLU (5-Shot): 63.94
- TruthfulQA (0-shot): 59.80
- Winogrande (5-shot): 79.64
Good For
- General-purpose text generation: Capable of producing coherent and contextually relevant text for various applications.
- Reasoning tasks: Its scores on reasoning benchmarks suggest suitability for tasks requiring logical inference.
- Instruction-tuned applications: Can be effectively used in conversational agents or systems requiring adherence to specific prompts.
- Developers seeking a balanced 7B model: Offers a strong blend of performance and efficiency for a model of its size.