arcee-ai/Hermes-Mistral-Saul-Slerp
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 7, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold
Hermes-Mistral-Saul-Slerp is a 7 billion parameter language model created by arcee-ai, built upon the Mistral architecture. This model is a merge of Equall/Saul-Instruct-v1 and NousResearch/Nous-Hermes-2-Mistral-7B-DPO using the slerp method. It is designed to combine the strengths of its constituent models, offering a versatile foundation for various natural language processing tasks.
Loading preview...
Model Overview
arcee-ai/Hermes-Mistral-Saul-Slerp is a 7 billion parameter language model based on the Mistral architecture, developed by arcee-ai. This model is a result of merging two distinct models: Equall/Saul-Instruct-v1 and NousResearch/Nous-Hermes-2-Mistral-7B-DPO.
Key Characteristics
- Merge Method: Utilizes the
slerp(spherical linear interpolation) merge method viamergekitto combine the weights of its base models. - Base Models: Integrates capabilities from both
Equall/Saul-Instruct-v1andNousResearch/Nous-Hermes-2-Mistral-7B-DPO. - Architecture: Inherits the efficient and performant Mistral architecture, providing a context length of 4096 tokens.
Good For
- General-purpose NLP: Suitable for a broad range of natural language understanding and generation tasks, leveraging the combined strengths of its merged components.
- Experimentation: Ideal for developers and researchers looking to explore the effects of model merging and the synergistic capabilities of different instruction-tuned models.