Brouz/Slerpeno

TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Sep 8, 2023License:cc-by-4.0Architecture:Transformer0.0K Open Weights Cold

Brouz/Slerpeno is a 13 billion parameter language model developed by Brouz, utilizing the same base models as Stheno but employing the SLERP merging method. This model is designed for general language tasks, offering a distinct approach to model fusion. Its 4096-token context length supports a range of applications requiring moderate input and output lengths.

Loading preview...

Overview

Brouz/Slerpeno is a 13 billion parameter language model that distinguishes itself through its unique merging methodology. Developed by Brouz, this model leverages the same foundational architectures as the Stheno series but integrates them using the Spherical Linear Interpolation (SLERP) method. This approach to model merging can yield different performance characteristics and emergent capabilities compared to other merging techniques.

Key Capabilities

  • 13 Billion Parameters: Offers a substantial capacity for understanding and generating complex language.
  • SLERP Merging: Utilizes Spherical Linear Interpolation for combining base models, potentially leading to unique and optimized performance profiles.
  • 4096-Token Context Length: Capable of processing and generating text within a moderate context window, suitable for various conversational and document-based tasks.

Good For

  • General Language Generation: Suitable for a wide array of text generation tasks, from creative writing to summarization.
  • Research into Model Merging: Provides a practical example of SLERP-based model fusion for developers and researchers interested in this technique.
  • Applications Requiring Moderate Context: Effective for use cases where the input and output text lengths fit within a 4096-token window.