FredrikBL/NeuralPipe-7B-slerp

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 6, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

FredrikBL/NeuralPipe-7B-slerp is a 7 billion parameter language model created by FredrikBL, formed by merging OpenPipe/mistral-ft-optimized-1218 and mlabonne/NeuralHermes-2.5-Mistral-7B using a slerp merge method. This model leverages the strengths of its base components, offering a balanced performance profile for general-purpose text generation tasks. It is designed for applications requiring a capable 7B model with a 4096-token context length.

Loading preview...

NeuralPipe-7B-slerp Overview

NeuralPipe-7B-slerp is a 7 billion parameter language model developed by FredrikBL. It is a product of a slerp merge (Spherical Linear Interpolation) of two distinct base models:

  • OpenPipe/mistral-ft-optimized-1218: A Mistral-based model optimized by OpenPipe.
  • mlabonne/NeuralHermes-2.5-Mistral-7B: Another Mistral-based model, NeuralHermes-2.5, known for its strong performance.

This merging technique combines the strengths of both parent models, aiming to create a more versatile and robust language model. The merge configuration specifically applies varying interpolation values (t) across different layers and attention mechanisms (self_attn) and multi-layer perceptrons (mlp) to fine-tune the resulting model's characteristics.

Key Capabilities

  • General-purpose text generation: Suitable for a wide range of conversational and creative tasks.
  • Leverages Mistral architecture: Benefits from the efficiency and performance of the Mistral 7B base.
  • Context length: Supports a standard context window of 4096 tokens.

Good for

  • Developers looking for a merged 7B model that combines the characteristics of two well-regarded Mistral variants.
  • Applications requiring a balance of performance and efficiency from a 7B parameter model.
  • Experimentation with models created via advanced merging techniques like slerp.