AurelPx/Pegasus-7b-slerp

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 22, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

AurelPx/Pegasus-7b-slerp is a 7 billion parameter language model created by AurelPx, formed by merging ammarali32/multi_verse_model and eren23/dpo-binarized-NeutrixOmnibe-7B using a slerp method. This model leverages the strengths of its constituent models to provide a versatile base for various natural language processing tasks. It is designed for general-purpose text generation and understanding within a 4096-token context window.

Loading preview...

Pegasus-7b-slerp: A Merged 7B Language Model

Pegasus-7b-slerp is a 7 billion parameter language model developed by AurelPx. It was created through a slerp (spherical linear interpolation) merge of two distinct base models: ammarali32/multi_verse_model and eren23/dpo-binarized-NeutrixOmnibe-7B. This merging technique allows for combining the learned representations of different models, potentially leading to improved performance across various tasks.

Key Capabilities

  • Model Merging: Utilizes the LazyMergekit framework with a slerp method to blend the weights of its parent models.
  • Versatile Base: Inherits capabilities from both multi_verse_model and dpo-binarized-NeutrixOmnibe-7B, making it suitable for a range of general-purpose text generation and comprehension tasks.
  • Standard Context Window: Supports a context length of 4096 tokens, allowing for processing moderately sized inputs.

When to Use This Model

This model is a good candidate for developers looking for a 7B parameter model that benefits from the combined strengths of multiple fine-tuned models. It can be used for applications requiring text generation, summarization, question answering, and other common NLP tasks where a balanced performance from merged architectures is desired.