Weyaxi/openchat-3.5-1210-Seraph-Slerp
Weyaxi/openchat-3.5-1210-Seraph-Slerp is a 7 billion parameter language model created by Weyaxi, merged using the slerp method from openchat/openchat-3.5-1210 and Weyaxi/Seraph-7B, based on Mistral-7B-v0.1. This model leverages a specific slerp merging strategy to combine the strengths of its base models, making it suitable for general conversational AI tasks. Its 4096-token context length supports moderate-length interactions.
Loading preview...
Model Overview
Weyaxi/openchat-3.5-1210-Seraph-Slerp is a 7 billion parameter language model developed by Weyaxi. It is a product of model merging, specifically utilizing the slerp (spherical linear interpolation) method via mergekit to combine two distinct base models: openchat/openchat-3.5-1210 and Weyaxi/Seraph-7B. The foundational architecture for this merged model is mistralai/Mistral-7B-v0.1.
Key Characteristics
- Merged Architecture: Combines
openchat-3.5-1210andSeraph-7Busing slerp interpolation. - Parameter-Specific Merging: The slerp method applies different interpolation ratios (
tvalues) to self-attention and MLP layers, allowing for fine-grained control over how the characteristics of the source models are blended. - Tokenizer: Uses a union tokenizer, ensuring compatibility with the vocabularies of its constituent models.
- Precision: Trained with
bfloat16dtype for efficient computation.
Potential Use Cases
- General Conversational AI: Suitable for chatbots and interactive applications, inheriting capabilities from its openchat base.
- Research in Model Merging: Provides an example of advanced slerp merging techniques for combining LLMs.