louisgrc/Montebello_7B_SLERP
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Mar 26, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

Montebello_7B_SLERP is a 7 billion parameter language model developed by louisgrc, created by merging yam-peleg/Experiment21-7B and louisgrc/Marengoli_7B_SLERP using the SLERP merge method. This model leverages the strengths of its constituent models to provide a balanced performance profile. It is suitable for general text generation tasks, offering a context length of 8192 tokens.

Loading preview...

Montebello_7B_SLERP Overview

Montebello_7B_SLERP is a 7 billion parameter language model developed by louisgrc, distinguished by its creation through a SLERP (Spherical Linear Interpolation) merge. This model is a composite of two distinct base models: yam-peleg/Experiment21-7B and louisgrc/Marengoli_7B_SLERP.

Key Characteristics

  • Merge Method: Utilizes the SLERP merging technique, which is known for smoothly interpolating between model weights, potentially leading to a more harmonious blend of capabilities from the source models.
  • Constituent Models: Built upon yam-peleg/Experiment21-7B and louisgrc/Marengoli_7B_SLERP, suggesting an aim to combine their respective strengths.
  • Parameter Configuration: The merge configuration specifies different interpolation values (t) for self-attention and MLP layers, indicating a fine-tuned approach to how each component model contributes to the final architecture.
  • Context Length: Supports an 8192-token context window, enabling it to handle moderately long inputs and generate coherent, extended responses.

Good For

  • General Text Generation: Suitable for a wide range of natural language processing tasks, including content creation, summarization, and conversational AI.
  • Experimentation with Merged Models: Offers a practical example of a model created via advanced merging techniques, useful for researchers and developers exploring model fusion.
  • Applications requiring a 7B model: Provides a capable option for scenarios where a 7 billion parameter model fits computational and performance requirements.
Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p