FelixChao/WestSeverus-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 23, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

FelixChao/WestSeverus-7B is a 7 billion parameter language model, created by merging senseable/WestLake-7B-v2 and FelixChao/Severus-7B using a slerp merge method. This model leverages the strengths of its constituent models, offering a balanced performance profile for general text generation tasks. With a 4096-token context length, it is suitable for applications requiring moderate context understanding and generation.

Loading preview...

WestSeverus-7B Overview

WestSeverus-7B is a 7 billion parameter language model developed by FelixChao, resulting from a strategic merge of two distinct models: senseable/WestLake-7B-v2 and FelixChao/Severus-7B. This integration was performed using the LazyMergekit tool, employing a slerp merge method.

Key Characteristics

  • Merged Architecture: Combines the strengths of WestLake-7B-v2 and Severus-7B, aiming for a synergistic performance.
  • Parameter Count: A 7 billion parameter model, offering a good balance between performance and computational efficiency.
  • Context Length: Supports a context window of 4096 tokens, enabling it to handle moderately long inputs and generate coherent responses.
  • Merge Configuration: The merge process specifically configured different parameter weights for self-attention and MLP layers, indicating a fine-tuned approach to combine the models effectively.

Ideal Use Cases

  • General Text Generation: Suitable for a wide array of tasks including content creation, summarization, and conversational AI.
  • Research and Experimentation: Provides a solid base for further fine-tuning or exploring merged model architectures.
  • Applications requiring moderate context: Its 4096-token context window makes it viable for tasks where understanding and generating text based on a few paragraphs of information is sufficient.