cris177/Orca-Hermes-7B-slerp

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 23, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

cris177/Orca-Hermes-7B-slerp is a 7 billion parameter language model created by cris177, merged from Open-Orca/Mistral-7B-OpenOrca and teknium/OpenHermes-2.5-Mistral-7B using the slerp method. This model leverages the strengths of both base models, offering a balanced performance profile for general-purpose language tasks. With a 4096-token context length, it is suitable for applications requiring moderate input and output lengths.

Loading preview...

Orca-Hermes-7B-slerp Overview

Orca-Hermes-7B-slerp is a 7 billion parameter language model developed by cris177, resulting from a strategic merge of two prominent Mistral-7B-based models: Open-Orca/Mistral-7B-OpenOrca and teknium/OpenHermes-2.5-Mistral-7B. This merge was performed using the slerp (spherical linear interpolation) method via mergekit, aiming to combine the distinct capabilities of its constituent models.

Key Characteristics

  • Merged Architecture: Combines the instruction-following and reasoning capabilities of Open-Orca with the conversational and creative strengths of OpenHermes-2.5.
  • Parameter Count: A 7 billion parameter model, offering a balance between performance and computational efficiency.
  • Context Length: Supports a context window of 4096 tokens, suitable for a variety of tasks requiring moderate input and output lengths.
  • Merge Method: Utilizes slerp for merging, which can lead to a more harmonious blend of model weights compared to simpler averaging methods.

Good For

  • General-purpose AI applications: Its merged nature makes it versatile for a wide range of tasks.
  • Instruction-following and conversational AI: Benefits from the fine-tuning of both Open-Orca and OpenHermes-2.5.
  • Developers seeking a balanced 7B model: Offers a potentially robust option without the need to train from scratch.