allknowingroger/M7merge-7B-slerp

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kLicense:apache-2.0Architecture:Transformer Open Weights Cold

M7merge-7B-slerp by allknowingroger is a 7 billion parameter language model created by merging automerger/M7T3qm7x-7B and automerger/T3qm7xpStrangemerges_32-7B using the slerp method. This model leverages a specific layer-wise parameter interpolation strategy to combine the strengths of its constituent models. It is designed for general text generation tasks, inheriting capabilities from its merged base models.

Loading preview...

Overview

M7merge-7B-slerp is a 7 billion parameter language model developed by allknowingroger. It is a product of merging two distinct models: automerger/M7T3qm7x-7B and automerger/T3qm7xpStrangemerges_32-7B. The merge was performed using the slerp (spherical linear interpolation) method, a technique often employed in model merging to combine parameters smoothly.

Key Characteristics

  • Merge Method: Utilizes slerp for combining model weights, specifically interpolating parameters across different layers.
  • Base Models: Constructed from automerger/M7T3qm7x-7B and automerger/T3qm7xpStrangemerges_32-7B, suggesting a blend of their respective capabilities.
  • Parameter Configuration: The merge applies varying t values for self_attn and mlp filters, indicating a nuanced approach to how different parts of the neural network are combined.
  • Data Type: Configured to use bfloat16 for model operations, balancing performance and memory efficiency.

Usage

This model can be loaded and used with the Hugging Face transformers library for text generation tasks. The provided configuration demonstrates how to apply a chat template and generate text with specified parameters like max_new_tokens, temperature, top_k, and top_p.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p