arcee-ai/Hermes-Mistral-Saul-Slerp

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 7, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Hermes-Mistral-Saul-Slerp is a 7 billion parameter language model created by arcee-ai, built upon the Mistral architecture. This model is a merge of Equall/Saul-Instruct-v1 and NousResearch/Nous-Hermes-2-Mistral-7B-DPO using the slerp method. It is designed to combine the strengths of its constituent models, offering a versatile foundation for various natural language processing tasks.

Loading preview...

Model Overview

arcee-ai/Hermes-Mistral-Saul-Slerp is a 7 billion parameter language model based on the Mistral architecture, developed by arcee-ai. This model is a result of merging two distinct models: Equall/Saul-Instruct-v1 and NousResearch/Nous-Hermes-2-Mistral-7B-DPO.

Key Characteristics

  • Merge Method: Utilizes the slerp (spherical linear interpolation) merge method via mergekit to combine the weights of its base models.
  • Base Models: Integrates capabilities from both Equall/Saul-Instruct-v1 and NousResearch/Nous-Hermes-2-Mistral-7B-DPO.
  • Architecture: Inherits the efficient and performant Mistral architecture, providing a context length of 4096 tokens.

Good For

  • General-purpose NLP: Suitable for a broad range of natural language understanding and generation tasks, leveraging the combined strengths of its merged components.
  • Experimentation: Ideal for developers and researchers looking to explore the effects of model merging and the synergistic capabilities of different instruction-tuned models.